var/home/core/zuul-output/0000755000175000017500000000000015067471703014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067506525015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004555741515067506516017730 0ustar rootrootOct 02 12:49:14 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 12:49:14 crc restorecon[4671]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:14 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 12:49:15 crc restorecon[4671]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 12:49:16 crc kubenswrapper[4710]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 12:49:16 crc kubenswrapper[4710]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 12:49:16 crc kubenswrapper[4710]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 12:49:16 crc kubenswrapper[4710]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 12:49:16 crc kubenswrapper[4710]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 12:49:16 crc kubenswrapper[4710]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.524944 4710 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530378 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530405 4710 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530412 4710 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530418 4710 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530424 4710 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530431 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530436 4710 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530442 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530449 4710 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530455 4710 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530464 4710 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530472 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530478 4710 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530485 4710 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530491 4710 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530498 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530504 4710 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530510 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530516 4710 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530522 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530527 4710 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530533 4710 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530539 4710 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530546 4710 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530553 4710 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530560 4710 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530566 4710 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530572 4710 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530577 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530583 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530589 4710 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530595 4710 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530601 4710 feature_gate.go:330] unrecognized feature gate: Example Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530606 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530612 4710 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530618 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530623 4710 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530629 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530634 4710 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530640 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530645 4710 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530650 4710 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530658 4710 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530665 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530670 4710 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530675 4710 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530680 4710 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530687 4710 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530693 4710 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530700 4710 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530706 4710 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530712 4710 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530717 4710 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530723 4710 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530728 4710 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530733 4710 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530738 4710 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530763 4710 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530769 4710 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530774 4710 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530780 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530785 4710 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530792 4710 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530798 4710 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530803 4710 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530808 4710 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530813 4710 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530818 4710 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530823 4710 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530829 4710 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.530834 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.530978 4710 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.530993 4710 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531009 4710 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531018 4710 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531028 4710 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531035 4710 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531044 4710 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531052 4710 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531059 4710 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531065 4710 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531072 4710 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531078 4710 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531084 4710 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531090 4710 flags.go:64] FLAG: --cgroup-root="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531096 4710 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531104 4710 flags.go:64] FLAG: --client-ca-file="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531110 4710 flags.go:64] FLAG: --cloud-config="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531116 4710 flags.go:64] FLAG: --cloud-provider="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531121 4710 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531129 4710 flags.go:64] FLAG: --cluster-domain="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531135 4710 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531141 4710 flags.go:64] FLAG: --config-dir="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531147 4710 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531153 4710 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531161 4710 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531168 4710 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531174 4710 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531180 4710 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531186 4710 flags.go:64] FLAG: --contention-profiling="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531192 4710 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531198 4710 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531204 4710 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531211 4710 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531218 4710 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531224 4710 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531231 4710 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531237 4710 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531243 4710 flags.go:64] FLAG: --enable-server="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.531249 4710 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533036 4710 flags.go:64] FLAG: --event-burst="100" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533049 4710 flags.go:64] FLAG: --event-qps="50" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533054 4710 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533058 4710 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533063 4710 flags.go:64] FLAG: --eviction-hard="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533069 4710 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533074 4710 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533078 4710 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533083 4710 flags.go:64] FLAG: --eviction-soft="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533087 4710 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533091 4710 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533123 4710 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533128 4710 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533133 4710 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533137 4710 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533141 4710 flags.go:64] FLAG: --feature-gates="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533147 4710 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533151 4710 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533155 4710 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533160 4710 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533164 4710 flags.go:64] FLAG: --healthz-port="10248" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533168 4710 flags.go:64] FLAG: --help="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533172 4710 flags.go:64] FLAG: --hostname-override="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533176 4710 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533181 4710 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533185 4710 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533188 4710 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533192 4710 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533197 4710 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533202 4710 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533206 4710 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533211 4710 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533215 4710 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533219 4710 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533223 4710 flags.go:64] FLAG: --kube-reserved="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533228 4710 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533233 4710 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533238 4710 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533243 4710 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533248 4710 flags.go:64] FLAG: --lock-file="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533253 4710 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533258 4710 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533263 4710 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533271 4710 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533276 4710 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533281 4710 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533287 4710 flags.go:64] FLAG: --logging-format="text" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533291 4710 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533297 4710 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533301 4710 flags.go:64] FLAG: --manifest-url="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533305 4710 flags.go:64] FLAG: --manifest-url-header="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533311 4710 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533315 4710 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533321 4710 flags.go:64] FLAG: --max-pods="110" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533326 4710 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533330 4710 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533334 4710 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533338 4710 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533342 4710 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533346 4710 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533352 4710 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533365 4710 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533370 4710 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533375 4710 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533379 4710 flags.go:64] FLAG: --pod-cidr="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533383 4710 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533389 4710 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533393 4710 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533398 4710 flags.go:64] FLAG: --pods-per-core="0" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533402 4710 flags.go:64] FLAG: --port="10250" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533407 4710 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533411 4710 flags.go:64] FLAG: --provider-id="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533414 4710 flags.go:64] FLAG: --qos-reserved="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533418 4710 flags.go:64] FLAG: --read-only-port="10255" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533423 4710 flags.go:64] FLAG: --register-node="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533427 4710 flags.go:64] FLAG: --register-schedulable="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533431 4710 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533438 4710 flags.go:64] FLAG: --registry-burst="10" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533442 4710 flags.go:64] FLAG: --registry-qps="5" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533446 4710 flags.go:64] FLAG: --reserved-cpus="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533450 4710 flags.go:64] FLAG: --reserved-memory="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533456 4710 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533459 4710 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533464 4710 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533468 4710 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533473 4710 flags.go:64] FLAG: --runonce="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533477 4710 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533481 4710 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533485 4710 flags.go:64] FLAG: --seccomp-default="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533490 4710 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533494 4710 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533499 4710 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533503 4710 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533507 4710 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533512 4710 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533516 4710 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533521 4710 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533525 4710 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533529 4710 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533533 4710 flags.go:64] FLAG: --system-cgroups="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533537 4710 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533544 4710 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533548 4710 flags.go:64] FLAG: --tls-cert-file="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533552 4710 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533557 4710 flags.go:64] FLAG: --tls-min-version="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533561 4710 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533565 4710 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533569 4710 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533573 4710 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533577 4710 flags.go:64] FLAG: --v="2" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533584 4710 flags.go:64] FLAG: --version="false" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533591 4710 flags.go:64] FLAG: --vmodule="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533595 4710 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.533600 4710 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535202 4710 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535220 4710 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535228 4710 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535234 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535245 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535251 4710 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535259 4710 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535266 4710 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535278 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535284 4710 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535289 4710 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535294 4710 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535299 4710 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535304 4710 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535309 4710 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535316 4710 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535374 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535492 4710 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.535578 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536110 4710 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536131 4710 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536150 4710 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536162 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536174 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536183 4710 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536191 4710 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536209 4710 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536220 4710 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536229 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536238 4710 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536246 4710 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536255 4710 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536263 4710 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536271 4710 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536280 4710 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536288 4710 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536303 4710 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536318 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536327 4710 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536339 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536348 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536359 4710 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536368 4710 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536377 4710 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536386 4710 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536394 4710 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536402 4710 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536410 4710 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536418 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536427 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536439 4710 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536453 4710 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536464 4710 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536475 4710 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536483 4710 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536493 4710 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536501 4710 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536515 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536524 4710 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536533 4710 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536543 4710 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536555 4710 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536565 4710 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536575 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536583 4710 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536592 4710 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536602 4710 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536610 4710 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536619 4710 feature_gate.go:330] unrecognized feature gate: Example Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536627 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.536635 4710 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.536654 4710 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.597417 4710 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.597486 4710 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597618 4710 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597634 4710 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597644 4710 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597658 4710 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597671 4710 feature_gate.go:330] unrecognized feature gate: Example Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597682 4710 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597691 4710 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597699 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597707 4710 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597715 4710 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597724 4710 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597732 4710 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597765 4710 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597774 4710 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597782 4710 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597790 4710 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597798 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597806 4710 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597814 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597821 4710 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597829 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597837 4710 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597845 4710 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597852 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597860 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597868 4710 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597877 4710 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597885 4710 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597894 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597901 4710 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597909 4710 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597916 4710 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597924 4710 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597931 4710 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597940 4710 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597950 4710 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597958 4710 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597966 4710 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597973 4710 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597981 4710 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597989 4710 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.597997 4710 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598008 4710 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598019 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598029 4710 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598039 4710 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598048 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598056 4710 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598066 4710 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598076 4710 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598086 4710 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598094 4710 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598103 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598111 4710 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598119 4710 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598127 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598134 4710 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598142 4710 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598150 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598158 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598166 4710 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598175 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598183 4710 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598192 4710 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598201 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598210 4710 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598218 4710 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598229 4710 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598237 4710 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598245 4710 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598253 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.598268 4710 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598502 4710 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598513 4710 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598522 4710 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598532 4710 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598541 4710 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598549 4710 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598556 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598564 4710 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598572 4710 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598580 4710 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598587 4710 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598595 4710 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598602 4710 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598611 4710 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598618 4710 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598626 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598634 4710 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598641 4710 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598650 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598660 4710 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598667 4710 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598675 4710 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598682 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598691 4710 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598699 4710 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598706 4710 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598714 4710 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598722 4710 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598731 4710 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598739 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598769 4710 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598777 4710 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598785 4710 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598792 4710 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598800 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598808 4710 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598815 4710 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598823 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598831 4710 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598839 4710 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598846 4710 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598857 4710 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598867 4710 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598877 4710 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598888 4710 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598899 4710 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598909 4710 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598917 4710 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598927 4710 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598935 4710 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598944 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598954 4710 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598964 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598972 4710 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598980 4710 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.598999 4710 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599007 4710 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599014 4710 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599022 4710 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599030 4710 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599038 4710 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599046 4710 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599054 4710 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599062 4710 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599070 4710 feature_gate.go:330] unrecognized feature gate: Example Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599078 4710 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599085 4710 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599093 4710 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599100 4710 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599108 4710 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.599115 4710 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.599128 4710 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.599435 4710 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.642404 4710 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.642566 4710 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.645089 4710 server.go:997] "Starting client certificate rotation" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.645144 4710 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.645364 4710 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-16 03:49:23.073534943 +0000 UTC Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.645567 4710 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1071h0m6.427971566s for next certificate rotation Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.694055 4710 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.697564 4710 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.723825 4710 log.go:25] "Validated CRI v1 runtime API" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.771378 4710 log.go:25] "Validated CRI v1 image API" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.774213 4710 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.781560 4710 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-12-43-51-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.781628 4710 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.803788 4710 manager.go:217] Machine: {Timestamp:2025-10-02 12:49:16.801097496 +0000 UTC m=+0.907508399 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:5d561d62-be46-4109-8ed7-b35f0572f5b0 BootID:da9dd993-7c58-4d65-9aab-f7f0f0784918 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:4c:46:cc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:4c:46:cc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:78:b9:d7 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bf:18:3d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:65:af:17 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:4d:09:77 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8e:53:08:0f:42:11 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5e:48:77:50:bc:b2 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.804041 4710 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.804257 4710 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.804620 4710 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.805031 4710 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.805074 4710 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.808068 4710 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.808117 4710 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.808989 4710 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.809034 4710 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.811378 4710 state_mem.go:36] "Initialized new in-memory state store" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.811578 4710 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.816090 4710 kubelet.go:418] "Attempting to sync node with API server" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.816148 4710 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.816363 4710 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.816391 4710 kubelet.go:324] "Adding apiserver pod source" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.816409 4710 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.822336 4710 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.823684 4710 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.825660 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.825826 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.825913 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.825960 4710 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.825983 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827467 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827493 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827501 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827508 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827521 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827530 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827539 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827552 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827563 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827576 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827589 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.827595 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.828918 4710 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.829577 4710 server.go:1280] "Started kubelet" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.829800 4710 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.831466 4710 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.831476 4710 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 12:49:16 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.836221 4710 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.839251 4710 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.839462 4710 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.839543 4710 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:39:48.340547538 +0000 UTC Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.839809 4710 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1131h50m31.500744745s for next certificate rotation Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.839896 4710 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.839906 4710 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.840070 4710 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.841538 4710 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.841580 4710 factory.go:55] Registering systemd factory Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.841591 4710 factory.go:221] Registration of the systemd container factory successfully Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.840982 4710 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.841947 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.842096 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.842249 4710 factory.go:153] Registering CRI-O factory Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.842278 4710 factory.go:221] Registration of the crio container factory successfully Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.842310 4710 factory.go:103] Registering Raw factory Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.842330 4710 manager.go:1196] Started watching for new ooms in manager Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.843322 4710 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="200ms" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.849200 4710 server.go:460] "Adding debug handlers to kubelet server" Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.850146 4710 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.103:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aad7d00510bdc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 12:49:16.8294287 +0000 UTC m=+0.935839573,LastTimestamp:2025-10-02 12:49:16.8294287 +0000 UTC m=+0.935839573,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.852610 4710 manager.go:319] Starting recovery of all containers Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.855945 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856027 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856046 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856061 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856076 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856090 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856104 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856118 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856134 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856147 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856160 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856174 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856188 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856205 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856219 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856234 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856246 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856260 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856274 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856291 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856307 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856318 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856331 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856345 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856357 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856371 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856388 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856403 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856439 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856453 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856485 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856499 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856514 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856527 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856540 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856553 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856566 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856615 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856632 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856647 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856666 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856681 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856696 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856708 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856722 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856736 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856784 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856798 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856811 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856850 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856866 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856879 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856909 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856922 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856937 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856952 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856969 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856982 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.856998 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857012 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857024 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857037 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857051 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857064 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857079 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857093 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857105 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857119 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857133 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857146 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857159 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857173 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857186 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857199 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857212 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857226 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857240 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857253 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857269 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857283 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857297 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857315 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857330 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857345 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857360 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857375 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.857388 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.860903 4710 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.860943 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.860961 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.860977 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.860992 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861010 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861024 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861039 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861054 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861068 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861085 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861100 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861115 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861135 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861153 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861169 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861187 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861204 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861230 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861250 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861269 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861287 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861303 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861319 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861338 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861353 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861368 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861401 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861417 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861435 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861449 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861464 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861480 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861494 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861509 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861523 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861537 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861553 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861567 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861580 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861594 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861609 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861635 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861649 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861662 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861678 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861693 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861706 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861721 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861735 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861848 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861862 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861882 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861896 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861912 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861926 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861941 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861956 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861970 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.861989 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862007 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862039 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862055 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862074 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862088 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862101 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862116 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862130 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862145 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862174 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862188 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862203 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862216 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862232 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862247 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862266 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862281 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862295 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862310 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862324 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862341 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862355 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862369 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862385 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862399 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862413 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862428 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862446 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862460 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862475 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862489 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862506 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862520 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862536 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862551 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862567 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862583 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862597 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862611 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862625 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862640 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862656 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862677 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862690 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862705 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862721 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862735 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862769 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862783 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862798 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862815 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862832 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862846 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862863 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862877 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862893 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862910 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862924 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862939 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862952 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862966 4710 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862979 4710 reconstruct.go:97] "Volume reconstruction finished" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.862990 4710 reconciler.go:26] "Reconciler: start to sync state" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.876265 4710 manager.go:324] Recovery completed Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.889124 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.891311 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.891395 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.891415 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.892467 4710 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.892486 4710 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.892505 4710 state_mem.go:36] "Initialized new in-memory state store" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.901773 4710 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.903106 4710 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.903158 4710 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.903188 4710 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.903242 4710 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 12:49:16 crc kubenswrapper[4710]: W1002 12:49:16.905343 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.905453 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.913326 4710 policy_none.go:49] "None policy: Start" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.914181 4710 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.914252 4710 state_mem.go:35] "Initializing new in-memory state store" Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.941840 4710 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.970614 4710 manager.go:334] "Starting Device Plugin manager" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.970678 4710 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.970696 4710 server.go:79] "Starting device plugin registration server" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.971188 4710 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.971211 4710 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.971445 4710 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.971528 4710 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 12:49:16 crc kubenswrapper[4710]: I1002 12:49:16.971535 4710 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 12:49:16 crc kubenswrapper[4710]: E1002 12:49:16.979326 4710 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.003965 4710 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.004102 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.005244 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.005298 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.005308 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.005457 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.005700 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.005832 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.006129 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.006149 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.006157 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.006275 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.006709 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.006760 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007054 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007073 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007080 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007161 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007361 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007428 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007516 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007560 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007571 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007900 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007949 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.007960 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008102 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008242 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008284 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008854 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008907 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008914 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008948 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008922 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.008950 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.009061 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.009088 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.009101 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.009336 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.009380 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.010070 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.010102 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.010118 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.010391 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.010431 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.010444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.049945 4710 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="400ms" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065217 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065318 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065360 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065385 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065413 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065440 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065464 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065550 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065597 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065658 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065676 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065711 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065732 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065778 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.065818 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.071594 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.073435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.073487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.073502 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.073554 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.074177 4710 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167191 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167300 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167349 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167387 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167423 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167457 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167490 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167526 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167542 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167635 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167549 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167633 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167563 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167693 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167798 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167849 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167724 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167793 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167808 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167939 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.167977 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168037 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168072 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168034 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168148 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168107 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168195 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168189 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168241 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.168215 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.274878 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.277711 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.277857 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.277892 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.277951 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.278904 4710 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.339554 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.351737 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.362049 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.396314 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-c77ae409e3f882fc48dd737240d6f2672ac9f51fe7c07170978d5606d54f7a94 WatchSource:0}: Error finding container c77ae409e3f882fc48dd737240d6f2672ac9f51fe7c07170978d5606d54f7a94: Status 404 returned error can't find the container with id c77ae409e3f882fc48dd737240d6f2672ac9f51fe7c07170978d5606d54f7a94 Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.398492 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5bb018efaa92986f8be7c4393960d42ac740c05303bd892a3faab1e73d9b91b8 WatchSource:0}: Error finding container 5bb018efaa92986f8be7c4393960d42ac740c05303bd892a3faab1e73d9b91b8: Status 404 returned error can't find the container with id 5bb018efaa92986f8be7c4393960d42ac740c05303bd892a3faab1e73d9b91b8 Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.400574 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.403147 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-00c4aafb4f5ca5bcddeee9645c60bbffa28f56255210ecf300e1ff16cca369cc WatchSource:0}: Error finding container 00c4aafb4f5ca5bcddeee9645c60bbffa28f56255210ecf300e1ff16cca369cc: Status 404 returned error can't find the container with id 00c4aafb4f5ca5bcddeee9645c60bbffa28f56255210ecf300e1ff16cca369cc Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.414282 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.451794 4710 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="800ms" Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.461501 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c4323507ae7c0bfbe25ed3bac54ff16aeeddb0c650e4f6e7c291b45aec0844ed WatchSource:0}: Error finding container c4323507ae7c0bfbe25ed3bac54ff16aeeddb0c650e4f6e7c291b45aec0844ed: Status 404 returned error can't find the container with id c4323507ae7c0bfbe25ed3bac54ff16aeeddb0c650e4f6e7c291b45aec0844ed Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.470630 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1a8e54bc32c90fc98ecabac3caa375968d907f7ad36b3fadc4e93105af2363b5 WatchSource:0}: Error finding container 1a8e54bc32c90fc98ecabac3caa375968d907f7ad36b3fadc4e93105af2363b5: Status 404 returned error can't find the container with id 1a8e54bc32c90fc98ecabac3caa375968d907f7ad36b3fadc4e93105af2363b5 Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.631740 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.631897 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.679413 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.680794 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.680856 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.680872 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.680911 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.681572 4710 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.688429 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.688553 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.745214 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.745346 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:17 crc kubenswrapper[4710]: W1002 12:49:17.817436 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:17 crc kubenswrapper[4710]: E1002 12:49:17.817582 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.831111 4710 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.908056 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5bb018efaa92986f8be7c4393960d42ac740c05303bd892a3faab1e73d9b91b8"} Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.910607 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c77ae409e3f882fc48dd737240d6f2672ac9f51fe7c07170978d5606d54f7a94"} Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.912606 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1a8e54bc32c90fc98ecabac3caa375968d907f7ad36b3fadc4e93105af2363b5"} Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.914190 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c4323507ae7c0bfbe25ed3bac54ff16aeeddb0c650e4f6e7c291b45aec0844ed"} Oct 02 12:49:17 crc kubenswrapper[4710]: I1002 12:49:17.916448 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"00c4aafb4f5ca5bcddeee9645c60bbffa28f56255210ecf300e1ff16cca369cc"} Oct 02 12:49:18 crc kubenswrapper[4710]: E1002 12:49:18.253564 4710 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="1.6s" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.482692 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.484934 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.484995 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.485011 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.485156 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:18 crc kubenswrapper[4710]: E1002 12:49:18.485987 4710 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.831103 4710 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.923544 4710 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664" exitCode=0 Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.923636 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.923770 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.925575 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.925615 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.925628 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.928654 4710 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1b67d047f7215fc4019b97c40d85cbc8da042a7ae24c0123d8aae771e54001aa" exitCode=0 Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.928701 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1b67d047f7215fc4019b97c40d85cbc8da042a7ae24c0123d8aae771e54001aa"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.928763 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.929731 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.929773 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.929784 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.932002 4710 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909" exitCode=0 Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.932105 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.932191 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.933626 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.933835 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.933873 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.935368 4710 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="dbdcb8eb2e49a7163e9eaa828e7b78acabb2ce6f90a7b68903ba319ce0a93bb4" exitCode=0 Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.935464 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.935464 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"dbdcb8eb2e49a7163e9eaa828e7b78acabb2ce6f90a7b68903ba319ce0a93bb4"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.938549 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.938969 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.939026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.939047 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.939632 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.939691 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.939712 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.941220 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.941278 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.941306 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.941330 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b"} Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.941332 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.957140 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.957181 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:18 crc kubenswrapper[4710]: I1002 12:49:18.957197 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:19 crc kubenswrapper[4710]: W1002 12:49:19.245230 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:19 crc kubenswrapper[4710]: E1002 12:49:19.245348 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:19 crc kubenswrapper[4710]: W1002 12:49:19.350043 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:19 crc kubenswrapper[4710]: E1002 12:49:19.350226 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:19 crc kubenswrapper[4710]: W1002 12:49:19.586291 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:19 crc kubenswrapper[4710]: E1002 12:49:19.586395 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.830808 4710 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:19 crc kubenswrapper[4710]: E1002 12:49:19.855034 4710 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="3.2s" Oct 02 12:49:19 crc kubenswrapper[4710]: W1002 12:49:19.874868 4710 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Oct 02 12:49:19 crc kubenswrapper[4710]: E1002 12:49:19.874972 4710 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.955006 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.955056 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.955079 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.955093 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.956069 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.956104 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.956114 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.957955 4710 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="93b539ff68fd93ec655e82797012aaa080fc07231fa3ef6f3c1b88acfc0e0707" exitCode=0 Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.958046 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"93b539ff68fd93ec655e82797012aaa080fc07231fa3ef6f3c1b88acfc0e0707"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.958076 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.958844 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.958873 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.958888 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.962970 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.963028 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.963044 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.963093 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.965534 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d15d2975fb8f2164c1db8b41748f097b2b42ed7341b154606385fe605700128a"} Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.965555 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.965603 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.968512 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.968596 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.968614 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.969261 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.969299 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:19 crc kubenswrapper[4710]: I1002 12:49:19.969336 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.086564 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.088125 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.088184 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.088195 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.088231 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:20 crc kubenswrapper[4710]: E1002 12:49:20.088924 4710 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.359476 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.971939 4710 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3667ad1e2308a18aca2ab0dea63a75fb7ce9e0df597355e9d3d1970478914e8e" exitCode=0 Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.972072 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3667ad1e2308a18aca2ab0dea63a75fb7ce9e0df597355e9d3d1970478914e8e"} Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.972238 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.973783 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.973863 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.973904 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.978863 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd"} Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.978981 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.979044 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.979078 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.979084 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.978987 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.980564 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.980643 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.980669 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981005 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981060 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981084 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981226 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981270 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981292 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.981934 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.982008 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:20 crc kubenswrapper[4710]: I1002 12:49:20.982034 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.642878 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.989107 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5d6ac973b7bf0def8737e4a683c5ec6671fc36e56ea22ccef44cf0b95bf8d52b"} Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.989180 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"287f7de4750403e36b4fd25ee98d245448e844fc1a06b8d6445bb6b54b575bb2"} Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.989203 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ac07842001fdbaaa75ddfe29fee32ad211a3c4efae7a00262562675161c76124"} Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.989250 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.990283 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.990315 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:21 crc kubenswrapper[4710]: I1002 12:49:21.990327 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.390544 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.390901 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.392659 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.392720 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.392741 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.399630 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.656585 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.666838 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.997507 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.997948 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2903d593fcfe6586b1a94b65c35cd4fa8644210f5c7aabd7988a2a1901996a1a"} Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.998004 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3b5ee9ba2cb46bdb9df24738b7d9d01638a487a11f0c1a6841abc77dee63a48e"} Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.998109 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.998180 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.999250 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.999296 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.999311 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.999780 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.999807 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:22 crc kubenswrapper[4710]: I1002 12:49:22.999815 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:22.999918 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:22.999961 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:22.999978 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.125026 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.125232 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.127611 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.127709 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.127733 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.289742 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.291423 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.291469 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.291483 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.291512 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:23 crc kubenswrapper[4710]: I1002 12:49:23.439163 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.000336 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.000430 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.000336 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002100 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002123 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002337 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002364 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002375 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002447 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002490 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:24 crc kubenswrapper[4710]: I1002 12:49:24.002505 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.003171 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.005161 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.005208 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.005219 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.516619 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.516919 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.518495 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.518557 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.518584 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.983584 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.983775 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.985220 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.985291 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:25 crc kubenswrapper[4710]: I1002 12:49:25.985312 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:26 crc kubenswrapper[4710]: E1002 12:49:26.979728 4710 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 12:49:28 crc kubenswrapper[4710]: I1002 12:49:28.984108 4710 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 12:49:28 crc kubenswrapper[4710]: I1002 12:49:28.984242 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 12:49:30 crc kubenswrapper[4710]: I1002 12:49:30.367184 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:30 crc kubenswrapper[4710]: I1002 12:49:30.367539 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:30 crc kubenswrapper[4710]: I1002 12:49:30.370706 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:30 crc kubenswrapper[4710]: I1002 12:49:30.370898 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:30 crc kubenswrapper[4710]: I1002 12:49:30.370993 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:30 crc kubenswrapper[4710]: I1002 12:49:30.832494 4710 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 12:49:31 crc kubenswrapper[4710]: I1002 12:49:31.814052 4710 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 12:49:31 crc kubenswrapper[4710]: I1002 12:49:31.814135 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 12:49:31 crc kubenswrapper[4710]: I1002 12:49:31.820451 4710 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 12:49:31 crc kubenswrapper[4710]: I1002 12:49:31.820534 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.292594 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.293003 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.295756 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.296085 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.296278 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.342748 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.405906 4710 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]log ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]etcd ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/generic-apiserver-start-informers ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/priority-and-fairness-filter ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-apiextensions-informers ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-apiextensions-controllers ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/crd-informer-synced ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-system-namespaces-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 02 12:49:32 crc kubenswrapper[4710]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 02 12:49:32 crc kubenswrapper[4710]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/bootstrap-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/start-kube-aggregator-informers ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-registration-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-discovery-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]autoregister-completion ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-openapi-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 02 12:49:32 crc kubenswrapper[4710]: livez check failed Oct 02 12:49:32 crc kubenswrapper[4710]: I1002 12:49:32.406653 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:49:33 crc kubenswrapper[4710]: I1002 12:49:33.025667 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:33 crc kubenswrapper[4710]: I1002 12:49:33.027248 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:33 crc kubenswrapper[4710]: I1002 12:49:33.027329 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:33 crc kubenswrapper[4710]: I1002 12:49:33.027345 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:33 crc kubenswrapper[4710]: I1002 12:49:33.048217 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 12:49:34 crc kubenswrapper[4710]: I1002 12:49:34.028362 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:34 crc kubenswrapper[4710]: I1002 12:49:34.030111 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:34 crc kubenswrapper[4710]: I1002 12:49:34.030175 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:34 crc kubenswrapper[4710]: I1002 12:49:34.030195 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.807557 4710 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.811084 4710 trace.go:236] Trace[1413054244]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 12:49:24.055) (total time: 12755ms): Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[1413054244]: ---"Objects listed" error: 12755ms (12:49:36.810) Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[1413054244]: [12.755783015s] [12.755783015s] END Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.811121 4710 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.812604 4710 trace.go:236] Trace[553878701]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 12:49:22.998) (total time: 13814ms): Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[553878701]: ---"Objects listed" error: 13814ms (12:49:36.812) Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[553878701]: [13.814245016s] [13.814245016s] END Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.812648 4710 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.813026 4710 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.813107 4710 trace.go:236] Trace[934418237]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 12:49:25.502) (total time: 11310ms): Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[934418237]: ---"Objects listed" error: 11310ms (12:49:36.812) Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[934418237]: [11.310693864s] [11.310693864s] END Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.813144 4710 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.813691 4710 trace.go:236] Trace[1040575704]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 12:49:23.913) (total time: 12900ms): Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[1040575704]: ---"Objects listed" error: 12900ms (12:49:36.813) Oct 02 12:49:36 crc kubenswrapper[4710]: Trace[1040575704]: [12.90038322s] [12.90038322s] END Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.813710 4710 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.815917 4710 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.829912 4710 apiserver.go:52] "Watching apiserver" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.836525 4710 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.837054 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.837699 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.837774 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.837918 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.838240 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.838246 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.838357 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.838411 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.838588 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.838683 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.841110 4710 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.841164 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.842417 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.843090 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.843892 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.844107 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.846462 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.847101 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.851767 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.852229 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.900199 4710 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43990->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.900274 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43990->192.168.126.11:17697: read: connection reset by peer" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.900372 4710 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46382->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.900389 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46382->192.168.126.11:17697: read: connection reset by peer" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913504 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913554 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913578 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913604 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913632 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913656 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913676 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913695 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913716 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913735 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913773 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913796 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913816 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913834 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913853 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913872 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913894 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913916 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913938 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913983 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914005 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914028 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914052 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914073 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914099 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914119 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914141 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914162 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914185 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914206 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914226 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914247 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914266 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914299 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914320 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914360 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914384 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914404 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914427 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914448 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914469 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914490 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914516 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914565 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914585 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914608 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914628 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914652 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914672 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914691 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914710 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914731 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914786 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914811 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914833 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914872 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914912 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914933 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914980 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915002 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915021 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915042 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915064 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915086 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915105 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915157 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915211 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915255 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915277 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915299 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915318 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915342 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915363 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915387 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915407 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915429 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915450 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915475 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915495 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915517 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915538 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915560 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915581 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915601 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915622 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915644 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915664 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915688 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915708 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915731 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915771 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915824 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915846 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915869 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915891 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915914 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915935 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915955 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915973 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916080 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916106 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916129 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916149 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916171 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916192 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916216 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916237 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916257 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.913977 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916279 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914048 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916301 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916325 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916346 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916369 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916391 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916413 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916436 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916458 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916479 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916503 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916526 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916547 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916572 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916596 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916618 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916639 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916661 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916682 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916703 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916726 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916764 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916786 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916808 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916828 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916848 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916870 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916891 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916945 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916971 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916994 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917016 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917037 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917058 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917079 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917101 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917124 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917148 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917169 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917189 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917211 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917234 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917255 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917275 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917295 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917319 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917341 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917362 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917384 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917405 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917427 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917448 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917468 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917491 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917513 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917536 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917559 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917579 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917599 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917626 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917650 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917675 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917697 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917719 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917745 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917791 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917815 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917840 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917862 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917885 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917908 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917935 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917984 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918008 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918032 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918059 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918085 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918110 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918133 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918157 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918179 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918201 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918222 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918248 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918295 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918324 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918352 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918402 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918427 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918449 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918474 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918503 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918532 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918560 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918587 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918610 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918635 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918658 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918706 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918721 4710 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919148 4710 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.926622 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914177 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914313 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914424 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.934549 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914585 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914686 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914863 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.914890 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915128 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915170 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915240 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915259 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915359 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915391 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915403 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915552 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915659 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915740 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915782 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915849 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.915915 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916005 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916190 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916251 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916266 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916378 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916410 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916437 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916563 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916677 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.916709 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917594 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917791 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917877 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.917995 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918316 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918512 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918561 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918665 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918707 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918805 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.918908 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919102 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919114 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919126 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919139 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919134 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.919360 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.919868 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:49:37.419844699 +0000 UTC m=+21.526255582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.920631 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.920842 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.921080 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.921088 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.920976 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.921396 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.921596 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.921772 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.921944 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922072 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922228 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922353 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922491 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922700 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922795 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922852 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.922919 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923100 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923118 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923150 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923384 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923393 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923424 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923705 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923761 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.923774 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.924044 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.924048 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.924134 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.924341 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.924562 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.925054 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.925233 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.925286 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.925663 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.925972 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.926245 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.926463 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.926511 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.926613 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.926884 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.930166 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.930453 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.932898 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.933200 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.933527 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.933770 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.934026 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.934311 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.934830 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935188 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935228 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935303 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935348 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935665 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935712 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935914 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.935954 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.936428 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.937184 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.937348 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.937698 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.937894 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.938205 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.938203 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.938600 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.939208 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.939253 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.939278 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.940702 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.940900 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.940992 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:37.440968561 +0000 UTC m=+21.547379564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.941328 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.941534 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.941575 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:37.441564336 +0000 UTC m=+21.547975319 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.941593 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.941855 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.943846 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.944124 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.944306 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.944643 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.945089 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.945392 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.945725 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.945872 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.947099 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.955750 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.956272 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.956373 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.956590 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.956725 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.956965 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.956967 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957085 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957388 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957573 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957690 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957704 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.958020 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.958063 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.958124 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.958262 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.958479 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.958482 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957633 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.957659 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.962918 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.962944 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.962959 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.963026 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:37.463009586 +0000 UTC m=+21.569420469 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.972605 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.972631 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.972677 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.972690 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:36 crc kubenswrapper[4710]: E1002 12:49:36.972766 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:37.472719426 +0000 UTC m=+21.579130309 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.973388 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.975526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.975901 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976200 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976323 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976382 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976487 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976533 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976787 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976797 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.976948 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.977180 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.977207 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.977224 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.977432 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.977535 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.977654 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.978165 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.979450 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.980023 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.980083 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.980154 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.981017 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.982003 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.982183 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.984119 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.984441 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.984525 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.991078 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.991477 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.994041 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.996319 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.997902 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.997977 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.998292 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.998433 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:36 crc kubenswrapper[4710]: I1002 12:49:36.999814 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.000356 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.000381 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.001208 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.003432 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.004121 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.004232 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.004276 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.004303 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.004446 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.004945 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.024264 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025117 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025250 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025406 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025424 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025441 4710 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025453 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025464 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025475 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025488 4710 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025498 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025511 4710 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025522 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025535 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025546 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025558 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025568 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025583 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025594 4710 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025604 4710 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025618 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025628 4710 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025639 4710 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025648 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025661 4710 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025671 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025682 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025693 4710 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025708 4710 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025718 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025729 4710 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025746 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025775 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025787 4710 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025799 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025813 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025825 4710 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025835 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025845 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025859 4710 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025870 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025882 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025893 4710 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025907 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025917 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025927 4710 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025942 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025952 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025962 4710 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025972 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025984 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.025995 4710 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026008 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026019 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026033 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026043 4710 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026056 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026066 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026080 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026090 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026101 4710 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026114 4710 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026123 4710 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026133 4710 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026142 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026155 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026165 4710 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026175 4710 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026186 4710 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026199 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026209 4710 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026221 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026234 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026243 4710 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026253 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026263 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026275 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026285 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026295 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026306 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026319 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026329 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026339 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026349 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026370 4710 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026380 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026390 4710 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026405 4710 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026416 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026425 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026435 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026449 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026459 4710 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026469 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026479 4710 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026492 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026503 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026521 4710 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026532 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026546 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026556 4710 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026569 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026584 4710 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026595 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026606 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026616 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026639 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026649 4710 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026660 4710 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026670 4710 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026684 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026694 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026704 4710 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026717 4710 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026727 4710 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026745 4710 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026772 4710 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026786 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026798 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026810 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026826 4710 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026839 4710 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026850 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026862 4710 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026872 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026895 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026906 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026916 4710 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026928 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026939 4710 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026949 4710 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026961 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026974 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026984 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.026994 4710 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027009 4710 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027023 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027034 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027045 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027058 4710 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027068 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027078 4710 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027088 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027100 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027111 4710 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027120 4710 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027132 4710 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027147 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027157 4710 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027167 4710 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027177 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027194 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027204 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027214 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027234 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027247 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027257 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027267 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027279 4710 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027289 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027299 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027327 4710 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027340 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027349 4710 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027359 4710 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027370 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027383 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027394 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027406 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027421 4710 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027430 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027440 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027450 4710 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027462 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027472 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027482 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027492 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027506 4710 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027517 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027527 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027538 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027546 4710 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027555 4710 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027563 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027574 4710 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027582 4710 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027591 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027603 4710 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027614 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027623 4710 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.027632 4710 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.028008 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.028020 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.028066 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.028912 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.035802 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.046080 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.054516 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.057219 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.065436 4710 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd" exitCode=255 Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.065496 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd"} Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.068118 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.079883 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.093523 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.102809 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.111595 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.127168 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.128463 4710 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.128482 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.139723 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.151324 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.158115 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.158396 4710 scope.go:117] "RemoveContainer" containerID="268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.160010 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.170500 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.177737 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.180215 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.334517 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.342919 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.350712 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.362279 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.372544 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.382229 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.391671 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.401845 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.403135 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.412126 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.429103 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.430575 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.430780 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:49:38.430729731 +0000 UTC m=+22.537140614 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.446134 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.456463 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.469093 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.490321 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.519621 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.531658 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.531701 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.531720 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.531739 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.531868 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.531882 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.531893 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.531941 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:38.531925663 +0000 UTC m=+22.638336546 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.531946 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.531996 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.532011 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.532080 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:38.532056716 +0000 UTC m=+22.638467599 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.532151 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.532182 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:38.532172139 +0000 UTC m=+22.638583122 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.532219 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: E1002 12:49:37.532244 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:38.532236101 +0000 UTC m=+22.638647084 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.542326 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.565444 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.579072 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.957356 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8ddjb"] Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.957787 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8ddjb" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.959968 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.963469 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.963656 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.963680 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.963867 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.969351 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-tblbf"] Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.969786 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-qpqlq"] Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.969989 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.970387 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-skmsh"] Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.970548 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.970780 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.979787 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.981732 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.981911 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.982937 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.983190 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.983223 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.983319 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.984649 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.984872 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.987257 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 12:49:37 crc kubenswrapper[4710]: I1002 12:49:37.998439 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.025800 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035012 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-cni-bin\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035075 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-etc-kubernetes\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035103 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km2kc\" (UniqueName: \"kubernetes.io/projected/7786a93b-a392-425e-b4ac-ed4956546d08-kube-api-access-km2kc\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035166 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-cni-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035192 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-socket-dir-parent\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035212 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-netns\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035234 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-os-release\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035253 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-kubelet\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035278 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-multus-certs\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035306 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-system-cni-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035330 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-k8s-cni-cncf-io\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035371 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-cnibin\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035447 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-cni-multus\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035472 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-conf-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035506 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7786a93b-a392-425e-b4ac-ed4956546d08-multus-daemon-config\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035552 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7786a93b-a392-425e-b4ac-ed4956546d08-cni-binary-copy\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.035575 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-hostroot\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.042831 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.055995 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.067530 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.069412 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.070490 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.071162 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.071309 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"98d46ac8b322aa35b3407ce65936808bca3b63f073e11c79ef0adf0a1894762c"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.072683 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.072811 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.072876 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ac5b2ad9617d8f671e1daab62843a9434d32426ccee3bc0ffe32ae4ee7fedcee"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.073730 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.073808 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e7ef1ae9c30d909b7221b99baec832ef9ca99780eabcd29cf7eae6bc4e32070e"} Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.076845 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.084205 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.098022 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.108349 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.122487 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136723 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-socket-dir-parent\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136779 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-netns\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136803 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01947451-6af9-4b65-b74c-87e098437b71-mcd-auth-proxy-config\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136829 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-multus-certs\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136845 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01947451-6af9-4b65-b74c-87e098437b71-proxy-tls\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136861 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cni-binary-copy\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136879 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-os-release\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136894 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-kubelet\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136907 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-k8s-cni-cncf-io\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136925 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-cni-multus\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136940 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7786a93b-a392-425e-b4ac-ed4956546d08-multus-daemon-config\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136931 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-socket-dir-parent\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.136955 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cnibin\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137027 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-multus-certs\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137060 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-k8s-cni-cncf-io\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137083 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-cni-multus\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137348 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-os-release\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137389 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-kubelet\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137417 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-run-netns\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137484 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-hostroot\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137516 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-system-cni-dir\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137547 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1f74797a-11a0-4174-b10d-0c6a63566f8a-hosts-file\") pod \"node-resolver-skmsh\" (UID: \"1f74797a-11a0-4174-b10d-0c6a63566f8a\") " pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137642 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-os-release\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137690 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-hostroot\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137708 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7786a93b-a392-425e-b4ac-ed4956546d08-multus-daemon-config\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137822 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-cni-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137855 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137875 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-system-cni-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137894 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-cnibin\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137911 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-conf-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137923 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-cni-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137935 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxw85\" (UniqueName: \"kubernetes.io/projected/1f74797a-11a0-4174-b10d-0c6a63566f8a-kube-api-access-kxw85\") pod \"node-resolver-skmsh\" (UID: \"1f74797a-11a0-4174-b10d-0c6a63566f8a\") " pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137971 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-system-cni-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137984 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-cnibin\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.137992 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhlsh\" (UniqueName: \"kubernetes.io/projected/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-kube-api-access-vhlsh\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138007 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-multus-conf-dir\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138024 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138089 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7786a93b-a392-425e-b4ac-ed4956546d08-cni-binary-copy\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138112 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/01947451-6af9-4b65-b74c-87e098437b71-rootfs\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138134 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6fbt\" (UniqueName: \"kubernetes.io/projected/01947451-6af9-4b65-b74c-87e098437b71-kube-api-access-w6fbt\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138159 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-cni-bin\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138187 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-etc-kubernetes\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138206 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km2kc\" (UniqueName: \"kubernetes.io/projected/7786a93b-a392-425e-b4ac-ed4956546d08-kube-api-access-km2kc\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138267 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-host-var-lib-cni-bin\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138307 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7786a93b-a392-425e-b4ac-ed4956546d08-etc-kubernetes\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.138541 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7786a93b-a392-425e-b4ac-ed4956546d08-cni-binary-copy\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.139683 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.156154 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.167075 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km2kc\" (UniqueName: \"kubernetes.io/projected/7786a93b-a392-425e-b4ac-ed4956546d08-kube-api-access-km2kc\") pod \"multus-8ddjb\" (UID: \"7786a93b-a392-425e-b4ac-ed4956546d08\") " pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.169395 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.184972 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.202154 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.214279 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.229494 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.238832 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cnibin\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.238894 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-system-cni-dir\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.238941 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1f74797a-11a0-4174-b10d-0c6a63566f8a-hosts-file\") pod \"node-resolver-skmsh\" (UID: \"1f74797a-11a0-4174-b10d-0c6a63566f8a\") " pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.238957 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-os-release\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.238963 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cnibin\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.238980 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239111 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-os-release\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239191 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxw85\" (UniqueName: \"kubernetes.io/projected/1f74797a-11a0-4174-b10d-0c6a63566f8a-kube-api-access-kxw85\") pod \"node-resolver-skmsh\" (UID: \"1f74797a-11a0-4174-b10d-0c6a63566f8a\") " pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239251 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhlsh\" (UniqueName: \"kubernetes.io/projected/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-kube-api-access-vhlsh\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239307 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/01947451-6af9-4b65-b74c-87e098437b71-rootfs\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239335 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6fbt\" (UniqueName: \"kubernetes.io/projected/01947451-6af9-4b65-b74c-87e098437b71-kube-api-access-w6fbt\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239356 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239413 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01947451-6af9-4b65-b74c-87e098437b71-mcd-auth-proxy-config\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239450 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01947451-6af9-4b65-b74c-87e098437b71-proxy-tls\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239488 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cni-binary-copy\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239569 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/01947451-6af9-4b65-b74c-87e098437b71-rootfs\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239245 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-system-cni-dir\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239067 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1f74797a-11a0-4174-b10d-0c6a63566f8a-hosts-file\") pod \"node-resolver-skmsh\" (UID: \"1f74797a-11a0-4174-b10d-0c6a63566f8a\") " pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.239885 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.240411 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01947451-6af9-4b65-b74c-87e098437b71-mcd-auth-proxy-config\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.240499 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cni-binary-copy\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.240669 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.247737 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.247830 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01947451-6af9-4b65-b74c-87e098437b71-proxy-tls\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.258271 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.261400 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxw85\" (UniqueName: \"kubernetes.io/projected/1f74797a-11a0-4174-b10d-0c6a63566f8a-kube-api-access-kxw85\") pod \"node-resolver-skmsh\" (UID: \"1f74797a-11a0-4174-b10d-0c6a63566f8a\") " pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.263285 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhlsh\" (UniqueName: \"kubernetes.io/projected/eda4d6c2-f0cd-43e6-b16e-44c5147e5281-kube-api-access-vhlsh\") pod \"multus-additional-cni-plugins-qpqlq\" (UID: \"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\") " pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.263946 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6fbt\" (UniqueName: \"kubernetes.io/projected/01947451-6af9-4b65-b74c-87e098437b71-kube-api-access-w6fbt\") pod \"machine-config-daemon-tblbf\" (UID: \"01947451-6af9-4b65-b74c-87e098437b71\") " pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.269992 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.272033 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8ddjb" Oct 02 12:49:38 crc kubenswrapper[4710]: W1002 12:49:38.281825 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7786a93b_a392_425e_b4ac_ed4956546d08.slice/crio-1a515a54a5dc1b4a7570a3ec36ae332d15fc75d251f93c1d4e4b62b144804e44 WatchSource:0}: Error finding container 1a515a54a5dc1b4a7570a3ec36ae332d15fc75d251f93c1d4e4b62b144804e44: Status 404 returned error can't find the container with id 1a515a54a5dc1b4a7570a3ec36ae332d15fc75d251f93c1d4e4b62b144804e44 Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.284854 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.285289 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.293567 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.301056 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-skmsh" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.313397 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 12:49:38 crc kubenswrapper[4710]: W1002 12:49:38.349920 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f74797a_11a0_4174_b10d_0c6a63566f8a.slice/crio-a4931fc7c7addcc91873624491eb24872963d473e5c49651f21e84936e7c6640 WatchSource:0}: Error finding container a4931fc7c7addcc91873624491eb24872963d473e5c49651f21e84936e7c6640: Status 404 returned error can't find the container with id a4931fc7c7addcc91873624491eb24872963d473e5c49651f21e84936e7c6640 Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.383632 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8jhhq"] Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.390702 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.395447 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.395608 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.395657 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.396135 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.396296 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.396477 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.396518 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.414120 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.432180 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.440280 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.440470 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:49:40.440438917 +0000 UTC m=+24.546849800 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.448508 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.466318 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.480407 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.495962 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.509268 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.523641 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.537281 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540724 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-systemd\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540780 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-etc-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540799 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-node-log\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540816 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-env-overrides\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540880 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlht4\" (UniqueName: \"kubernetes.io/projected/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-kube-api-access-wlht4\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540940 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-slash\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.540978 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-bin\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541001 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-script-lib\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541029 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-kubelet\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541048 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-ovn\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541088 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541112 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-netd\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541139 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-systemd-units\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541159 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-var-lib-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541193 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-netns\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541208 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541231 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovn-node-metrics-cert\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541261 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:40.54124309 +0000 UTC m=+24.647653973 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541299 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541323 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541353 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541379 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-config\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541400 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541428 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-ovn-kubernetes\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541453 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-log-socket\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.541500 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541537 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541552 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541563 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541587 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:40.541579598 +0000 UTC m=+24.647990471 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541691 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541715 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541727 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541702 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541803 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:40.541773883 +0000 UTC m=+24.648184766 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.541886 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:40.541850995 +0000 UTC m=+24.648261878 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.552845 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.576157 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.595521 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.608262 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642480 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642536 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-config\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642563 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-ovn-kubernetes\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642666 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642734 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-log-socket\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642596 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-log-socket\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642688 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-ovn-kubernetes\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642871 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-node-log\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642941 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-node-log\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.642896 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-systemd\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643045 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-systemd\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643083 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-etc-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643106 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-slash\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643153 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-etc-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643189 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-env-overrides\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643247 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-slash\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643500 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-config\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643876 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-env-overrides\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643934 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlht4\" (UniqueName: \"kubernetes.io/projected/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-kube-api-access-wlht4\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.643981 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-bin\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644062 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-bin\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644095 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-script-lib\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644125 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-kubelet\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644231 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-ovn\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644179 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-kubelet\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644321 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-netd\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644389 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-var-lib-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644430 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-netd\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644475 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-systemd-units\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644555 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-netns\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644594 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovn-node-metrics-cert\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644398 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-ovn\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644651 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-netns\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644481 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-var-lib-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644691 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-script-lib\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644514 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-systemd-units\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644706 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.644791 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-openvswitch\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.649438 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovn-node-metrics-cert\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.667050 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlht4\" (UniqueName: \"kubernetes.io/projected/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-kube-api-access-wlht4\") pod \"ovnkube-node-8jhhq\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.708578 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:38 crc kubenswrapper[4710]: W1002 12:49:38.725650 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-26968094c576b9eb7c131a4671b065d80104a9deef4b43b8bffe0bc631d967b8 WatchSource:0}: Error finding container 26968094c576b9eb7c131a4671b065d80104a9deef4b43b8bffe0bc631d967b8: Status 404 returned error can't find the container with id 26968094c576b9eb7c131a4671b065d80104a9deef4b43b8bffe0bc631d967b8 Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.905997 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.906152 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.906575 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.906635 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.906681 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:38 crc kubenswrapper[4710]: E1002 12:49:38.906721 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.909567 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.910421 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.911328 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.912135 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.912935 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.913629 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.914394 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.916620 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.917712 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.919117 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.919839 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.921119 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.921986 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.922845 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.923337 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.924245 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.924798 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.925168 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.926295 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.926864 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.927314 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.928257 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.928659 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.929659 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.930099 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.931069 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.931848 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.932654 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.933231 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.934176 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.934731 4710 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.935046 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.936842 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.937722 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.938352 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.939745 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.940717 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.941258 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.942249 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.943044 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.943888 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.944452 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.945451 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.946110 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.946902 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.947424 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.948252 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.948961 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.949775 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.950204 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.951003 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.951534 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.952124 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 12:49:38 crc kubenswrapper[4710]: I1002 12:49:38.953182 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.077610 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-skmsh" event={"ID":"1f74797a-11a0-4174-b10d-0c6a63566f8a","Type":"ContainerStarted","Data":"a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.077669 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-skmsh" event={"ID":"1f74797a-11a0-4174-b10d-0c6a63566f8a","Type":"ContainerStarted","Data":"a4931fc7c7addcc91873624491eb24872963d473e5c49651f21e84936e7c6640"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.079152 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerStarted","Data":"589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.079201 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerStarted","Data":"65fcf9ec658fdaf9d50320855e25bacabc7fc8b1e24009b9d7e9800e4d3fc3c9"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.082019 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1" exitCode=0 Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.082128 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.082166 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"26968094c576b9eb7c131a4671b065d80104a9deef4b43b8bffe0bc631d967b8"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.087473 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.087717 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.088129 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"1fff863fa8208cad53f5ff16eb2a93da439edf5fc0164807950c0cc73e9a9c47"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.092004 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8ddjb" event={"ID":"7786a93b-a392-425e-b4ac-ed4956546d08","Type":"ContainerStarted","Data":"b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.092075 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8ddjb" event={"ID":"7786a93b-a392-425e-b4ac-ed4956546d08","Type":"ContainerStarted","Data":"1a515a54a5dc1b4a7570a3ec36ae332d15fc75d251f93c1d4e4b62b144804e44"} Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.109511 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.143332 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.163138 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.177979 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.191470 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.206769 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.218943 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.235992 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.250923 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.264084 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.279117 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.290938 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.308234 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.323423 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.341028 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.356524 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.371440 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.385709 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.401039 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.414645 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.432184 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.445736 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.457675 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.472977 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.489422 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.511241 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.901073 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zmkll"] Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.901928 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.906787 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.908265 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.908497 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.908647 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.921151 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.938840 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.961898 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.976362 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:39 crc kubenswrapper[4710]: I1002 12:49:39.992824 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.009854 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.021506 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.035728 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.048977 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.058255 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d709a39d-880c-4499-8429-cd1c815af683-host\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.059096 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d709a39d-880c-4499-8429-cd1c815af683-serviceca\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.060003 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psbwn\" (UniqueName: \"kubernetes.io/projected/d709a39d-880c-4499-8429-cd1c815af683-kube-api-access-psbwn\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.074773 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.097647 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.099597 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.099763 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.099871 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.100155 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.100255 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.100356 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.101115 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.102630 4710 generic.go:334] "Generic (PLEG): container finished" podID="eda4d6c2-f0cd-43e6-b16e-44c5147e5281" containerID="589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe" exitCode=0 Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.102774 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerDied","Data":"589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe"} Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.121475 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.138273 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.152148 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.161304 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psbwn\" (UniqueName: \"kubernetes.io/projected/d709a39d-880c-4499-8429-cd1c815af683-kube-api-access-psbwn\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.161714 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d709a39d-880c-4499-8429-cd1c815af683-host\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.161755 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d709a39d-880c-4499-8429-cd1c815af683-serviceca\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.161810 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d709a39d-880c-4499-8429-cd1c815af683-host\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.162919 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d709a39d-880c-4499-8429-cd1c815af683-serviceca\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.164852 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.182237 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.182652 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psbwn\" (UniqueName: \"kubernetes.io/projected/d709a39d-880c-4499-8429-cd1c815af683-kube-api-access-psbwn\") pod \"node-ca-zmkll\" (UID: \"d709a39d-880c-4499-8429-cd1c815af683\") " pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.199256 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.210850 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.214394 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zmkll" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.238635 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.280232 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.332171 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.378172 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.418218 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.445338 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.468483 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.468646 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:49:44.468621215 +0000 UTC m=+28.575032098 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.476414 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.515686 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.558778 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.569661 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.569727 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.569774 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.569801 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569839 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569872 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569903 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569920 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569928 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:44.569902719 +0000 UTC m=+28.676313602 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569934 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569949 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:44.5699406 +0000 UTC m=+28.676351483 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.569969 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:44.569959 +0000 UTC m=+28.676369883 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.570044 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.570093 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.570110 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.570194 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:44.570168966 +0000 UTC m=+28.676579889 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.603058 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.904521 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.904521 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.904680 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:40 crc kubenswrapper[4710]: I1002 12:49:40.904521 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.904893 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:40 crc kubenswrapper[4710]: E1002 12:49:40.905093 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.108995 4710 generic.go:334] "Generic (PLEG): container finished" podID="eda4d6c2-f0cd-43e6-b16e-44c5147e5281" containerID="7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053" exitCode=0 Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.109084 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerDied","Data":"7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053"} Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.111577 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zmkll" event={"ID":"d709a39d-880c-4499-8429-cd1c815af683","Type":"ContainerStarted","Data":"72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7"} Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.111609 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zmkll" event={"ID":"d709a39d-880c-4499-8429-cd1c815af683","Type":"ContainerStarted","Data":"cbcff31d1e735535a07b58d289049dc99bbf323e74e374ea498efbe1744b50e6"} Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.123092 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.137428 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.150557 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.166607 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.180607 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.196535 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.207795 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.222164 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.236511 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.255152 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.264700 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.277796 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.290364 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.306786 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.319774 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.332820 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.351043 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.363574 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.378142 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.397413 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.435711 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.481813 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.521635 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.559063 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.602051 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.641523 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.679702 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:41 crc kubenswrapper[4710]: I1002 12:49:41.715675 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.122656 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37"} Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.126275 4710 generic.go:334] "Generic (PLEG): container finished" podID="eda4d6c2-f0cd-43e6-b16e-44c5147e5281" containerID="0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9" exitCode=0 Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.126312 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerDied","Data":"0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9"} Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.148686 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.161329 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.173133 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.190297 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.206038 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.221910 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.232949 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.242537 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.255852 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.268599 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.279046 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.333255 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.352193 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.388715 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:42Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.903698 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:42 crc kubenswrapper[4710]: E1002 12:49:42.903888 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.904082 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:42 crc kubenswrapper[4710]: E1002 12:49:42.904228 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:42 crc kubenswrapper[4710]: I1002 12:49:42.904388 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:42 crc kubenswrapper[4710]: E1002 12:49:42.904491 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.131152 4710 generic.go:334] "Generic (PLEG): container finished" podID="eda4d6c2-f0cd-43e6-b16e-44c5147e5281" containerID="3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814" exitCode=0 Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.131204 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerDied","Data":"3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.146982 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.161149 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.179086 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.191576 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.209422 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.216399 4710 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.218396 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.218447 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.218475 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.218605 4710 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.224243 4710 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.224488 4710 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.225800 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.225832 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.225843 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.225863 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.225878 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.226660 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.239361 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: E1002 12:49:43.244067 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.249442 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.249504 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.249518 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.249555 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.249570 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.253814 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: E1002 12:49:43.262046 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.267308 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.267347 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.267357 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.267377 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.267391 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.269488 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: E1002 12:49:43.283644 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.289026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.289113 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.289129 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.289147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.289159 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.291138 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: E1002 12:49:43.304594 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.308136 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.308699 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.308801 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.308867 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.308938 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.308999 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.321609 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: E1002 12:49:43.322705 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: E1002 12:49:43.322845 4710 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.328091 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.328156 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.328189 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.328215 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.328230 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.340100 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.359130 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:43Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.430352 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.430407 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.430422 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.430456 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.430472 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.533619 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.533662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.533674 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.533693 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.533705 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.635947 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.635999 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.636016 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.636050 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.636067 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.738848 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.738896 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.738908 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.738926 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.738941 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.840972 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.841002 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.841010 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.841026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.841036 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.943440 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.943483 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.943491 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.943508 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:43 crc kubenswrapper[4710]: I1002 12:49:43.943521 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:43Z","lastTransitionTime":"2025-10-02T12:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.046728 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.046826 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.046851 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.046875 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.046892 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.139798 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerStarted","Data":"7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.149501 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.149580 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.149602 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.149631 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.149650 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.159311 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.182080 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.195897 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.210286 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.230343 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.245963 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.252817 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.252862 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.252871 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.252888 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.252899 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.261802 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.277374 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.295921 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.322479 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.342118 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.355541 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.356196 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.356222 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.356230 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.356248 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.356258 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.373145 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.387886 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:44Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.460131 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.460186 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.460199 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.460219 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.460233 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.510824 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.511084 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:49:52.511061547 +0000 UTC m=+36.617472430 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.562659 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.562724 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.562742 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.562802 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.562820 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.612324 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.612387 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.612431 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.612466 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612506 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612598 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612718 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612767 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612783 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612602 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:52.612580077 +0000 UTC m=+36.718990960 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612842 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:52.612820553 +0000 UTC m=+36.719231536 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612861 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:52.612854164 +0000 UTC m=+36.719265187 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.612996 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.613043 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.613059 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.613154 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:52.61312663 +0000 UTC m=+36.719537603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.666059 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.666130 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.666145 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.666169 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.666185 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.769246 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.769291 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.769303 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.769320 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.769336 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.872263 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.872322 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.872334 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.872353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.872366 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.903737 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.903851 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.903956 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.904068 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.904286 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:44 crc kubenswrapper[4710]: E1002 12:49:44.904453 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.975482 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.975538 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.975560 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.975588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:44 crc kubenswrapper[4710]: I1002 12:49:44.975606 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:44Z","lastTransitionTime":"2025-10-02T12:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.078943 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.078998 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.079013 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.079038 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.079054 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.148204 4710 generic.go:334] "Generic (PLEG): container finished" podID="eda4d6c2-f0cd-43e6-b16e-44c5147e5281" containerID="7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474" exitCode=0 Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.148326 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerDied","Data":"7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.157200 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.157973 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.158035 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.168416 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.183213 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.183277 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.183299 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.183328 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.183353 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.186804 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.225035 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.234053 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.236123 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.237321 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.249437 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.264005 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.284896 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.285267 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.285297 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.285308 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.285325 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.285337 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.301119 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.315833 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.329523 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.342067 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.359687 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.374932 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.386677 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.389383 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.389439 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.389451 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.389475 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.389495 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.402441 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.420907 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.442566 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.457448 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.470459 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.488803 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.492294 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.492323 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.492333 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.492348 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.492357 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.516999 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.535408 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.556931 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.572687 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.588363 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.594681 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.594714 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.594727 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.594765 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.594781 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.607308 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.622789 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.632779 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.696713 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.696756 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.696767 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.696780 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.696789 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.799330 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.799371 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.799381 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.799397 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.799410 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.902573 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.902630 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.902642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.902662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:45 crc kubenswrapper[4710]: I1002 12:49:45.902676 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:45Z","lastTransitionTime":"2025-10-02T12:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.005006 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.005067 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.005083 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.005107 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.005121 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.107847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.107934 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.107967 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.108177 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.108344 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.183945 4710 generic.go:334] "Generic (PLEG): container finished" podID="eda4d6c2-f0cd-43e6-b16e-44c5147e5281" containerID="8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8" exitCode=0 Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.184053 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerDied","Data":"8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.184114 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.200285 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.210968 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.211083 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.211121 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.211136 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.211159 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.211173 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.224671 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.236892 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.253373 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.267388 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.279502 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.292360 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.303778 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.314082 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.314138 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.314151 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.314169 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.314181 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.323666 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.344107 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.359823 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.381052 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.395961 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.417778 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.417823 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.417837 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.417855 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.417868 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.520544 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.520595 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.520607 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.520626 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.520640 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.624016 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.624077 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.624094 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.624119 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.624138 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.727548 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.727664 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.727721 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.727804 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.727831 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.831364 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.831472 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.831495 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.831563 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.831591 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.904463 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.904518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.904518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:46 crc kubenswrapper[4710]: E1002 12:49:46.904668 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:46 crc kubenswrapper[4710]: E1002 12:49:46.904805 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:46 crc kubenswrapper[4710]: E1002 12:49:46.904953 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.923262 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.935817 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.935869 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.935886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.935911 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.935932 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:46Z","lastTransitionTime":"2025-10-02T12:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.939253 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.958846 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:46 crc kubenswrapper[4710]: I1002 12:49:46.978877 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.003866 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:46Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.028586 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.040565 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.040641 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.040654 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.040686 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.040718 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.047871 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.072885 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.092197 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.124371 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.137618 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.143216 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.143245 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.143256 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.143271 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.143280 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.146668 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.158159 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.173881 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:47Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.193049 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" event={"ID":"eda4d6c2-f0cd-43e6-b16e-44c5147e5281","Type":"ContainerStarted","Data":"9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.193933 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.245435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.245499 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.245519 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.245546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.245565 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.348943 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.349013 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.349031 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.349057 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.349075 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.452218 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.452265 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.452301 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.452324 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.452339 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.555086 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.555132 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.555144 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.555163 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.555177 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.658251 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.658529 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.658606 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.658670 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.658724 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.761516 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.761654 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.761673 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.761700 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.761717 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.864968 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.865057 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.865090 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.865121 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.865143 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.968302 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.968367 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.968390 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.968421 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:47 crc kubenswrapper[4710]: I1002 12:49:47.968443 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:47Z","lastTransitionTime":"2025-10-02T12:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.071247 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.071320 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.071346 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.071376 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.071395 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.174846 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.174913 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.174944 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.174972 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.174989 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.215654 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.234823 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.253681 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.278807 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.278891 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.278909 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.278944 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.278968 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.283199 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.302024 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.318627 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.333862 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.348693 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.360545 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.374877 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.381742 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.381813 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.381831 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.381857 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.381878 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.388335 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.404833 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.419977 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.436896 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.485035 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.485089 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.485106 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.485130 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.485145 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.589188 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.589270 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.589298 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.589331 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.589353 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.692589 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.692704 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.692729 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.692794 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.692816 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.795837 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.796419 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.796444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.796474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.796501 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.898960 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.899044 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.899068 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.899101 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.899125 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:48Z","lastTransitionTime":"2025-10-02T12:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.904318 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.904361 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:48 crc kubenswrapper[4710]: E1002 12:49:48.904477 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:48 crc kubenswrapper[4710]: I1002 12:49:48.904567 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:48 crc kubenswrapper[4710]: E1002 12:49:48.904616 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:48 crc kubenswrapper[4710]: E1002 12:49:48.904825 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.002716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.003093 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.003297 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.003513 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.003707 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.107476 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.107908 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.108101 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.108255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.108393 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.211723 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.211828 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.211847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.211870 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.211888 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.314732 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.315616 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.315818 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.315971 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.316011 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.419074 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.419129 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.419147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.419169 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.419186 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.521297 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.521349 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.521367 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.521386 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.521400 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.624647 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.624823 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.624856 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.624905 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.624930 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.728124 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.728188 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.728211 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.728241 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.728264 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.831098 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.831162 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.831174 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.831224 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.831237 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.933340 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.933402 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.933435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.933461 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:49 crc kubenswrapper[4710]: I1002 12:49:49.933478 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:49Z","lastTransitionTime":"2025-10-02T12:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.036553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.036602 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.036614 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.036632 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.036645 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.139365 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.139420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.139436 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.139459 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.139474 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.242133 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.242176 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.242184 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.242199 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.242210 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.356818 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.356875 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.356887 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.356908 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.356921 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.458863 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.458909 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.458921 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.458940 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.458954 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.514615 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4"] Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.515220 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.517052 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.517319 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.532074 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.543540 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.555459 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.561120 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.561166 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.561178 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.561196 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.561207 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.567356 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.579581 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.591136 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.603040 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.617056 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.631476 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.655921 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.663901 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.663988 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.664016 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.664052 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.664078 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.675215 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.677801 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plq59\" (UniqueName: \"kubernetes.io/projected/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-kube-api-access-plq59\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.677879 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.677899 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.677935 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.690387 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.729850 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.757185 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.766277 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.766307 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.766315 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.766330 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.766338 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.778478 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.778525 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.778548 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.778598 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plq59\" (UniqueName: \"kubernetes.io/projected/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-kube-api-access-plq59\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.779279 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.779330 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.779489 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:50Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.783327 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.819061 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plq59\" (UniqueName: \"kubernetes.io/projected/9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe-kube-api-access-plq59\") pod \"ovnkube-control-plane-749d76644c-nwkn4\" (UID: \"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.829939 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" Oct 02 12:49:50 crc kubenswrapper[4710]: W1002 12:49:50.842391 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aed137d_9ef0_4e9b_b9d2_b6dca3e2defe.slice/crio-423d0b8fd07e433784671d2738934bef82b21e3334d6af40334c32cffb85c61c WatchSource:0}: Error finding container 423d0b8fd07e433784671d2738934bef82b21e3334d6af40334c32cffb85c61c: Status 404 returned error can't find the container with id 423d0b8fd07e433784671d2738934bef82b21e3334d6af40334c32cffb85c61c Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.869302 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.869340 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.869349 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.869366 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.869377 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.903895 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.903913 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.903913 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:50 crc kubenswrapper[4710]: E1002 12:49:50.904170 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:50 crc kubenswrapper[4710]: E1002 12:49:50.904048 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:50 crc kubenswrapper[4710]: E1002 12:49:50.904270 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.972568 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.972609 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.972619 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.972636 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:50 crc kubenswrapper[4710]: I1002 12:49:50.972649 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:50Z","lastTransitionTime":"2025-10-02T12:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.075713 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.075789 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.075799 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.075821 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.075830 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.177368 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.177407 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.177420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.177439 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.177453 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.209052 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/0.log" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.212271 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6" exitCode=1 Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.212357 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.213173 4710 scope.go:117] "RemoveContainer" containerID="8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.214613 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" event={"ID":"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe","Type":"ContainerStarted","Data":"cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.214657 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" event={"ID":"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe","Type":"ContainerStarted","Data":"3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.214667 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" event={"ID":"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe","Type":"ContainerStarted","Data":"423d0b8fd07e433784671d2738934bef82b21e3334d6af40334c32cffb85c61c"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.228768 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.244185 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.262522 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.279892 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.279934 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.279946 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.279966 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.279991 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.284412 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.303811 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"vent handler 7\\\\nI1002 12:49:50.605406 5962 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:49:50.605475 5962 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605643 5962 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605734 5962 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605818 5962 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605874 5962 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605964 5962 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606179 5962 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606448 5962 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.324511 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.336330 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.348026 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.364403 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.375888 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.383676 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.383701 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.383713 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.383728 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.383738 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.388560 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.402439 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.416099 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.427213 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.437927 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.449667 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.459193 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.470029 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.482880 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.485993 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.486040 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.486056 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.486078 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.486096 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.502297 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"vent handler 7\\\\nI1002 12:49:50.605406 5962 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:49:50.605475 5962 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605643 5962 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605734 5962 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605818 5962 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605874 5962 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605964 5962 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606179 5962 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606448 5962 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.516715 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.530434 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.540951 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.551870 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.565607 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.582563 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.588353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.588393 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.588405 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.588425 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.588437 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.599385 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.613303 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.625030 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.639822 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.647225 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.667283 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.681392 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.691093 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.691135 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.691147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.691166 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.691177 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.700599 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"vent handler 7\\\\nI1002 12:49:50.605406 5962 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:49:50.605475 5962 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605643 5962 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605734 5962 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605818 5962 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605874 5962 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605964 5962 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606179 5962 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606448 5962 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.711699 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.725307 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.740634 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.753518 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.765063 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.778927 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.793551 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.793587 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.793597 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.793613 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.793625 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.795260 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.817632 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.829567 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.842450 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.853601 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.863765 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.896057 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.896097 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.896107 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.896119 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.896134 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.963292 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zqx98"] Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.963713 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:51 crc kubenswrapper[4710]: E1002 12:49:51.963782 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.979135 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.987256 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhd24\" (UniqueName: \"kubernetes.io/projected/d77362b1-cb9c-498f-8e21-9114ce2de0d3-kube-api-access-fhd24\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.987318 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.991426 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:51Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.998893 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.998952 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.998970 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.998994 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:51 crc kubenswrapper[4710]: I1002 12:49:51.999011 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:51Z","lastTransitionTime":"2025-10-02T12:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.006005 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.021497 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.040342 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"vent handler 7\\\\nI1002 12:49:50.605406 5962 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:49:50.605475 5962 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605643 5962 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605734 5962 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605818 5962 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605874 5962 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605964 5962 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606179 5962 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606448 5962 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.054282 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.067406 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.081173 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.088863 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhd24\" (UniqueName: \"kubernetes.io/projected/d77362b1-cb9c-498f-8e21-9114ce2de0d3-kube-api-access-fhd24\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.088955 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.089126 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.089206 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:52.589188201 +0000 UTC m=+36.695599104 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.099680 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.101510 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.101544 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.101554 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.101569 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.101579 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.111801 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.118880 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhd24\" (UniqueName: \"kubernetes.io/projected/d77362b1-cb9c-498f-8e21-9114ce2de0d3-kube-api-access-fhd24\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.123657 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.136647 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.148624 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.160977 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.171805 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.183606 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.203441 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.203479 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.203487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.203501 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.203511 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.221340 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/1.log" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.222037 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/0.log" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.225662 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8" exitCode=1 Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.225707 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.225831 4710 scope.go:117] "RemoveContainer" containerID="8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.226988 4710 scope.go:117] "RemoveContainer" containerID="a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.227313 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.245714 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.267116 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"vent handler 7\\\\nI1002 12:49:50.605406 5962 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:49:50.605475 5962 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605643 5962 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605734 5962 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605818 5962 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605874 5962 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605964 5962 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606179 5962 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606448 5962 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.286393 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.302446 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.306662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.306699 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.306713 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.306732 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.306768 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.319128 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.340528 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.353960 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.369637 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.386769 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.402704 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.409818 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.409861 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.409873 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.409925 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.409939 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.417964 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.434232 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.447035 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.462659 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.473899 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.489899 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:52Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.538946 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.538995 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.539004 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.539031 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.539040 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.593821 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.593955 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.594102 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:08.594074404 +0000 UTC m=+52.700485287 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.594235 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.594286 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:53.594275789 +0000 UTC m=+37.700686782 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.641487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.641553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.641571 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.641595 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.641612 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.701965 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.702018 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.702040 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.702067 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702156 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702249 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702249 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702302 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702322 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702172 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702382 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702367 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:08.70228821 +0000 UTC m=+52.808699163 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702366 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702452 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:08.702426613 +0000 UTC m=+52.808837526 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702478 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:08.702466264 +0000 UTC m=+52.808877287 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.702498 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:08.702488445 +0000 UTC m=+52.808899468 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.743863 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.743919 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.743936 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.743960 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.743976 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.846203 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.846258 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.846271 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.846288 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.846300 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.904547 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.904616 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.904733 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.904808 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.904900 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:52 crc kubenswrapper[4710]: E1002 12:49:52.904957 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.949219 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.949293 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.949314 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.949337 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:52 crc kubenswrapper[4710]: I1002 12:49:52.949353 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:52Z","lastTransitionTime":"2025-10-02T12:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.052961 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.053051 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.053081 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.053114 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.053135 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.156078 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.156109 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.156117 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.156130 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.156138 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.233040 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/1.log" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.258591 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.258673 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.258688 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.258708 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.258721 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.362628 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.362774 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.362800 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.362835 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.362858 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.364561 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.364616 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.364635 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.364662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.364678 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.384787 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:53Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.389696 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.389726 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.389738 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.389782 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.389796 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.410068 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:53Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.416257 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.416326 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.416340 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.416360 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.416374 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.431878 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:53Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.437304 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.437345 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.437362 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.437386 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.437401 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.455152 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:53Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.459981 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.460042 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.460059 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.460082 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.460099 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.471971 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:53Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.472086 4710 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.473618 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.473650 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.473659 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.473674 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.473685 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.576535 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.576591 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.576599 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.576614 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.576625 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.612725 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.613003 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.613151 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:55.613121411 +0000 UTC m=+39.719532334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.679294 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.679355 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.679372 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.679396 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.679416 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.783461 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.783536 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.783562 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.783593 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.783617 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.886487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.886548 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.886571 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.886599 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.886621 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.904171 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:53 crc kubenswrapper[4710]: E1002 12:49:53.904366 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.989198 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.989257 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.989274 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.989299 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:53 crc kubenswrapper[4710]: I1002 12:49:53.989318 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:53Z","lastTransitionTime":"2025-10-02T12:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.092499 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.092553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.092569 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.092593 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.092610 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.194488 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.194530 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.194540 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.194585 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.194623 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.296949 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.297014 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.297032 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.297057 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.297074 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.399821 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.399878 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.399899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.399921 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.399934 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.503229 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.503295 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.503319 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.503344 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.503363 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.605982 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.606126 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.606147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.606174 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.606191 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.709842 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.709902 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.709949 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.709975 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.709994 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.812705 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.812795 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.812814 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.812843 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.812862 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.903967 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.904034 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:54 crc kubenswrapper[4710]: E1002 12:49:54.904107 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.903982 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:54 crc kubenswrapper[4710]: E1002 12:49:54.904357 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:54 crc kubenswrapper[4710]: E1002 12:49:54.904511 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.915787 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.915826 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.915839 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.915854 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:54 crc kubenswrapper[4710]: I1002 12:49:54.915865 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:54Z","lastTransitionTime":"2025-10-02T12:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.018884 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.018939 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.018948 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.018961 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.018971 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.122434 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.122498 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.122515 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.122540 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.122556 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.226130 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.226221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.226246 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.226280 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.226303 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.328568 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.328606 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.328615 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.328629 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.328638 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.432218 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.432291 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.432308 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.432347 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.432364 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.535602 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.535975 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.535998 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.536028 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.536051 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.635424 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:55 crc kubenswrapper[4710]: E1002 12:49:55.635556 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:55 crc kubenswrapper[4710]: E1002 12:49:55.635620 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:49:59.635603007 +0000 UTC m=+43.742013890 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.637838 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.637865 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.637875 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.637890 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.637901 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.740492 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.740588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.740611 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.740642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.740671 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.844024 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.844074 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.844091 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.844114 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.844130 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.904169 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:55 crc kubenswrapper[4710]: E1002 12:49:55.904651 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.947024 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.947365 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.947546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.947738 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:55 crc kubenswrapper[4710]: I1002 12:49:55.947971 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:55Z","lastTransitionTime":"2025-10-02T12:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.049936 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.049997 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.050020 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.050050 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.050075 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.153825 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.153895 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.153918 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.153949 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.153970 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.256613 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.256676 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.256699 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.256729 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.256799 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.359889 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.359962 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.359986 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.360013 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.360029 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.463187 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.463239 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.463255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.463274 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.463287 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.566599 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.566678 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.566702 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.566735 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.566804 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.669615 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.669674 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.669691 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.669715 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.669732 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.773272 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.773313 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.773322 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.773338 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.773346 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.876079 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.876113 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.876123 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.876139 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.876151 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.903910 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.903959 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.903910 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:56 crc kubenswrapper[4710]: E1002 12:49:56.904073 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:56 crc kubenswrapper[4710]: E1002 12:49:56.904239 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:56 crc kubenswrapper[4710]: E1002 12:49:56.904372 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.917059 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.931260 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.955774 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a50f1918188a262d487ae088ba3370100ed2d6a1851567bfa14742130ebfee6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"message\\\":\\\"vent handler 7\\\\nI1002 12:49:50.605406 5962 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:49:50.605475 5962 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605643 5962 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 12:49:50.605734 5962 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605818 5962 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605874 5962 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.605964 5962 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606179 5962 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:49:50.606448 5962 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.971032 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.978304 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.978529 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.978630 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.978739 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.978862 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:56Z","lastTransitionTime":"2025-10-02T12:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.984137 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:56 crc kubenswrapper[4710]: I1002 12:49:56.996769 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.005326 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.018620 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.031889 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.043265 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.054643 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.067003 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.077677 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.081091 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.081132 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.081143 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.081158 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.081169 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.093522 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.107344 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.118233 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.183361 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.183389 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.183442 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.183485 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.183495 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.285921 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.285991 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.286009 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.286033 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.286051 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.388843 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.388937 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.388963 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.388991 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.389017 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.492803 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.493471 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.493508 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.493535 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.493553 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.596466 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.596548 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.596572 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.596603 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.596628 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.700034 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.700114 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.700138 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.700164 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.700181 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.803529 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.803935 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.804122 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.804306 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.804520 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.904033 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:57 crc kubenswrapper[4710]: E1002 12:49:57.904299 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.906681 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.906780 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.906811 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.906838 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:57 crc kubenswrapper[4710]: I1002 12:49:57.906861 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:57Z","lastTransitionTime":"2025-10-02T12:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.010197 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.010265 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.010283 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.010307 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.010319 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.113910 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.113992 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.114012 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.114036 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.114055 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.216081 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.217119 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.217151 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.217180 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.217194 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.217207 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.217362 4710 scope.go:117] "RemoveContainer" containerID="a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8" Oct 02 12:49:58 crc kubenswrapper[4710]: E1002 12:49:58.217711 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.242604 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.260878 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.278882 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.299820 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.319893 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.319963 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.319987 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.320018 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.320043 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.324162 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.340524 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.357590 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.373611 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.389989 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.411282 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.423907 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.423978 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.424004 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.424034 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.424057 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.433452 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.449776 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.463025 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.478331 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.492017 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.504526 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.531390 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.531437 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.531449 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.531465 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.531477 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.634168 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.634224 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.634242 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.634268 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.634287 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.737002 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.737031 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.737041 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.737055 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.737064 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.839508 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.839549 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.839558 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.839571 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.839579 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.904604 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.904682 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:49:58 crc kubenswrapper[4710]: E1002 12:49:58.904726 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.904801 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:49:58 crc kubenswrapper[4710]: E1002 12:49:58.904946 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:49:58 crc kubenswrapper[4710]: E1002 12:49:58.905060 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.942381 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.942445 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.942462 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.942485 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:58 crc kubenswrapper[4710]: I1002 12:49:58.942502 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:58Z","lastTransitionTime":"2025-10-02T12:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.045414 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.045472 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.045490 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.045513 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.045531 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.148868 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.148920 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.148936 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.148959 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.148979 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.252657 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.252709 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.252725 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.252763 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.252778 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.355121 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.355161 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.355169 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.355223 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.355236 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.457618 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.457678 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.457696 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.457715 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.457727 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.560112 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.560142 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.560151 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.560164 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.560173 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.662856 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.662927 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.662949 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.662980 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.663006 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.683929 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:59 crc kubenswrapper[4710]: E1002 12:49:59.684146 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:59 crc kubenswrapper[4710]: E1002 12:49:59.684266 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:07.684232121 +0000 UTC m=+51.790643044 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.766268 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.766489 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.766529 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.766564 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.766585 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.869486 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.869610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.869635 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.869666 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.869690 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.903407 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:49:59 crc kubenswrapper[4710]: E1002 12:49:59.903546 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.972813 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.972877 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.972888 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.972905 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:49:59 crc kubenswrapper[4710]: I1002 12:49:59.972917 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:49:59Z","lastTransitionTime":"2025-10-02T12:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.076383 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.076446 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.076459 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.076474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.076483 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.179787 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.179838 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.179854 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.179878 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.179897 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.282664 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.282704 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.282715 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.282732 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.282777 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.385503 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.385553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.385564 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.385583 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.385598 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.488120 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.488188 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.488205 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.488231 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.488249 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.591349 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.591404 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.591417 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.591435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.591448 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.694578 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.694640 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.694658 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.694681 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.694698 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.797420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.797482 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.797500 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.797524 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.797542 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.904296 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:00 crc kubenswrapper[4710]: E1002 12:50:00.904508 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.906391 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:00 crc kubenswrapper[4710]: E1002 12:50:00.906550 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.906828 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:00 crc kubenswrapper[4710]: E1002 12:50:00.906974 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.910435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.910486 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.910507 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.910537 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:00 crc kubenswrapper[4710]: I1002 12:50:00.910559 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:00Z","lastTransitionTime":"2025-10-02T12:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.013152 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.013211 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.013232 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.013259 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.013280 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.116256 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.116315 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.116331 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.116357 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.116374 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.219384 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.219459 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.219482 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.219512 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.219535 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.323521 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.323588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.323610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.323636 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.323655 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.427915 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.427964 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.427979 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.427999 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.428014 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.531400 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.531444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.531453 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.531467 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.531475 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.634716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.634835 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.634858 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.634890 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.634916 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.738599 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.738655 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.738671 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.738694 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.738710 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.842042 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.842098 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.842116 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.842167 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.842185 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.903719 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:01 crc kubenswrapper[4710]: E1002 12:50:01.903989 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.945067 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.945118 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.945127 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.945142 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:01 crc kubenswrapper[4710]: I1002 12:50:01.945152 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:01Z","lastTransitionTime":"2025-10-02T12:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.048085 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.048137 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.048160 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.048190 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.048213 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.151406 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.151474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.151496 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.151524 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.151546 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.255260 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.255331 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.255353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.255381 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.255403 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.358552 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.358600 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.358614 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.358630 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.358641 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.461528 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.461593 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.461610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.461639 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.461656 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.565142 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.565231 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.565244 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.565270 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.565287 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.668501 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.668538 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.668549 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.668566 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.668577 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.771153 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.771202 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.771210 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.771233 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.771246 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.874542 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.874598 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.874610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.874631 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.874648 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.904713 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.904904 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:02 crc kubenswrapper[4710]: E1002 12:50:02.904991 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.905021 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:02 crc kubenswrapper[4710]: E1002 12:50:02.905127 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:02 crc kubenswrapper[4710]: E1002 12:50:02.905233 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.978032 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.978116 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.978134 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.978159 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:02 crc kubenswrapper[4710]: I1002 12:50:02.978176 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:02Z","lastTransitionTime":"2025-10-02T12:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.081575 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.081618 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.081627 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.081653 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.081665 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.184883 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.185235 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.185366 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.185519 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.185633 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.288783 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.288820 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.288830 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.288847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.288857 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.391032 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.391077 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.391090 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.391110 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.391123 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.494167 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.494208 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.494221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.494239 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.494251 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.597882 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.597926 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.597938 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.598003 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.598017 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.606694 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.607342 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.607376 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.607395 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.607406 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.628359 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:03Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.632420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.632468 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.632480 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.632499 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.632511 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.647426 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:03Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.651417 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.651473 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.651486 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.651508 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.651521 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.668535 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:03Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.673345 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.673394 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.673406 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.673426 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.673440 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.688062 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:03Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.692594 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.692644 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.692659 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.692680 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.692697 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.705315 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:03Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.705434 4710 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.707072 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.707111 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.707126 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.707143 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.707155 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.810508 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.810571 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.810588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.810612 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.810629 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.904119 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:03 crc kubenswrapper[4710]: E1002 12:50:03.904330 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.912829 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.912888 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.912905 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.912928 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:03 crc kubenswrapper[4710]: I1002 12:50:03.912943 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:03Z","lastTransitionTime":"2025-10-02T12:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.015965 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.016025 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.016041 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.016066 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.016084 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.119317 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.119372 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.119389 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.119412 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.119429 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.222336 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.222382 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.222392 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.222407 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.222418 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.325079 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.325121 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.325132 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.325149 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.325161 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.427415 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.427479 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.427500 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.427525 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.427542 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.529925 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.529987 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.530000 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.530016 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.530028 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.632616 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.632654 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.632665 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.632681 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.632691 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.734798 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.734875 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.734897 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.734931 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.734954 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.838049 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.838143 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.838190 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.838218 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.838237 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.904208 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.904275 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.904322 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:04 crc kubenswrapper[4710]: E1002 12:50:04.904396 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:04 crc kubenswrapper[4710]: E1002 12:50:04.904486 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:04 crc kubenswrapper[4710]: E1002 12:50:04.904623 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.941588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.941620 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.941628 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.941642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:04 crc kubenswrapper[4710]: I1002 12:50:04.941650 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:04Z","lastTransitionTime":"2025-10-02T12:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.044410 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.044461 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.044473 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.044493 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.044505 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.146710 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.146839 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.146861 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.146887 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.146905 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.250205 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.250264 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.250281 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.250306 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.250323 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.353435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.353493 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.353509 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.353532 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.353549 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.456404 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.456451 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.456462 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.456479 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.456490 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.559152 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.559219 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.559236 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.559258 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.559273 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.662565 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.662610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.662625 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.662646 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.662660 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.765898 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.765966 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.765985 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.766010 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.766033 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.869584 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.869652 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.869678 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.869703 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.869720 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.904513 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:05 crc kubenswrapper[4710]: E1002 12:50:05.904907 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.973056 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.973111 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.973128 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.973152 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:05 crc kubenswrapper[4710]: I1002 12:50:05.973169 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:05Z","lastTransitionTime":"2025-10-02T12:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.076614 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.076677 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.076701 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.076732 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.076795 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.180141 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.180221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.180240 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.180263 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.180281 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.283127 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.283201 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.283221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.283249 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.283268 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.386960 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.387378 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.387613 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.387880 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.388092 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.491894 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.492264 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.492543 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.492820 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.493041 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.596685 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.597093 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.597241 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.597389 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.597538 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.700989 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.701067 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.701102 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.701134 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.701155 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.804656 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.804780 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.804796 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.804818 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.804833 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.904364 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.904415 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:06 crc kubenswrapper[4710]: E1002 12:50:06.904583 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.904634 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:06 crc kubenswrapper[4710]: E1002 12:50:06.904723 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:06 crc kubenswrapper[4710]: E1002 12:50:06.904864 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.908517 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.908804 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.909066 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.909263 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.909442 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:06Z","lastTransitionTime":"2025-10-02T12:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.921430 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.937498 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.949910 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.963576 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:06 crc kubenswrapper[4710]: I1002 12:50:06.982188 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.011789 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.013136 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.013189 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.013203 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.013231 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.013249 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.028176 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.044808 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.057087 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.076501 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.096393 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.113321 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.116157 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.116216 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.116234 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.116260 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.116278 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.127338 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.142933 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.164063 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.188102 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:07Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.219173 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.219247 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.219277 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.219296 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.219307 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.322198 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.322252 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.322263 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.322286 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.322298 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.425424 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.425534 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.425545 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.425564 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.425576 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.533405 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.533733 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.533886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.534026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.534150 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.636377 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.636433 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.636451 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.636477 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.636496 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.738915 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.738959 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.738970 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.738988 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.739000 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.777125 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:07 crc kubenswrapper[4710]: E1002 12:50:07.777336 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:50:07 crc kubenswrapper[4710]: E1002 12:50:07.777456 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:23.77741894 +0000 UTC m=+67.883829863 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.842096 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.842262 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.842275 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.842294 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.842307 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.904279 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:07 crc kubenswrapper[4710]: E1002 12:50:07.904493 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.944374 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.944430 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.944448 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.944472 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:07 crc kubenswrapper[4710]: I1002 12:50:07.944489 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:07Z","lastTransitionTime":"2025-10-02T12:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.047596 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.047657 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.047675 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.047701 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.047718 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.150779 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.150913 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.150933 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.150964 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.150983 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.253881 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.253921 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.253966 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.253984 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.253996 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.356761 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.356811 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.356821 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.356840 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.356854 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.460514 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.460575 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.460598 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.460629 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.460649 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.563688 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.563728 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.563738 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.563766 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.563778 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.666954 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.667012 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.667029 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.667051 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.667068 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.687660 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.687835 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:40.687814429 +0000 UTC m=+84.794225312 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.769602 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.769664 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.769684 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.769708 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.769726 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.788270 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.788381 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788426 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.788443 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788523 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:40.788493869 +0000 UTC m=+84.894904792 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.788581 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788596 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788644 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788641 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788664 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788682 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788706 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788740 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:40.788715764 +0000 UTC m=+84.895126677 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788819 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788843 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:40.788827257 +0000 UTC m=+84.895238180 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.788893 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:40.788871698 +0000 UTC m=+84.895282681 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.872954 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.873038 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.873065 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.873098 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.873124 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.904255 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.904289 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.904381 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.904501 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.904644 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:08 crc kubenswrapper[4710]: E1002 12:50:08.904725 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.975835 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.975876 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.975892 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.975916 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:08 crc kubenswrapper[4710]: I1002 12:50:08.975933 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:08Z","lastTransitionTime":"2025-10-02T12:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.079063 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.079130 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.079147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.079171 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.079190 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.182119 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.182181 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.182206 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.182232 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.182249 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.284845 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.284907 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.284923 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.284948 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.284965 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.388171 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.388238 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.388255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.388283 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.388300 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.491161 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.491204 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.491215 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.491230 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.491243 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.594632 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.594688 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.594698 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.594717 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.594728 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.697848 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.697917 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.697935 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.697963 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.697980 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.801097 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.801147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.801156 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.801174 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.801185 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.904060 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:09 crc kubenswrapper[4710]: E1002 12:50:09.904336 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.906162 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.906216 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.906236 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.906287 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:09 crc kubenswrapper[4710]: I1002 12:50:09.906308 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:09Z","lastTransitionTime":"2025-10-02T12:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.010207 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.010500 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.010585 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.010666 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.010961 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.113854 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.113909 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.113925 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.113946 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.113961 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.216869 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.216928 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.216946 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.216966 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.216981 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.319886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.319936 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.319946 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.319958 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.319967 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.423358 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.423420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.423438 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.423461 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.423479 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.526178 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.526214 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.526224 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.526239 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.526248 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.629175 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.629565 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.629577 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.629593 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.629602 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.733250 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.733309 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.733332 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.733360 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.733380 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.837139 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.837200 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.837217 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.837242 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.837260 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.904488 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.904618 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:10 crc kubenswrapper[4710]: E1002 12:50:10.904736 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.904885 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:10 crc kubenswrapper[4710]: E1002 12:50:10.904934 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:10 crc kubenswrapper[4710]: E1002 12:50:10.905132 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.939603 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.939842 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.940124 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.940264 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:10 crc kubenswrapper[4710]: I1002 12:50:10.940381 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:10Z","lastTransitionTime":"2025-10-02T12:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.043467 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.043520 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.043539 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.043564 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.043584 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.146539 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.146606 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.146629 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.146657 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.146677 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.250158 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.250221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.250238 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.250261 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.250280 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.353153 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.353194 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.353208 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.353229 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.353247 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.456440 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.456915 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.457101 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.457312 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.457467 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.560833 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.561417 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.561611 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.561827 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.562002 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.664363 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.664407 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.664418 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.664436 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.664449 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.767319 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.767357 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.767371 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.767391 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.767402 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.870175 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.870236 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.870254 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.870276 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.870295 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.904169 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:11 crc kubenswrapper[4710]: E1002 12:50:11.904365 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.905878 4710 scope.go:117] "RemoveContainer" containerID="a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.973776 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.973820 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.973834 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.973850 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:11 crc kubenswrapper[4710]: I1002 12:50:11.973861 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:11Z","lastTransitionTime":"2025-10-02T12:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.076491 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.076865 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.077006 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.077148 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.077330 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.181204 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.181525 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.181713 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.181998 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.182203 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.285281 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.285359 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.285383 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.285420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.285443 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.322358 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/1.log" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.325805 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.326465 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.341936 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.360387 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.388568 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.388642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.388657 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.388702 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.388717 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.396730 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.416186 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.438349 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.451139 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.466116 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.480605 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.491140 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.491348 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.491359 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.491377 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.491389 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.504612 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.520027 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.537773 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.551215 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.564050 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.578967 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.591998 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.593892 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.593930 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.593943 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.593960 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.593971 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.601952 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:12Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.695980 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.696010 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.696019 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.696035 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.696044 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.798004 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.798288 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.798297 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.798311 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.798322 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.901371 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.901404 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.901412 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.901426 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.901435 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:12Z","lastTransitionTime":"2025-10-02T12:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.903862 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.903902 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:12 crc kubenswrapper[4710]: I1002 12:50:12.903971 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:12 crc kubenswrapper[4710]: E1002 12:50:12.904034 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:12 crc kubenswrapper[4710]: E1002 12:50:12.904140 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:12 crc kubenswrapper[4710]: E1002 12:50:12.904259 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.003506 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.003563 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.003580 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.003606 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.003625 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.105634 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.105662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.105670 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.105683 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.105691 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.129599 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.141406 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.147811 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.159621 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.178639 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.202731 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.208663 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.208720 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.208785 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.208837 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.208862 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.227901 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.241217 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.257126 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.273471 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.292473 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.308425 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.310933 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.310988 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.311014 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.311044 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.311066 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.322815 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.331267 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/2.log" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.331985 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/1.log" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.339651 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a" exitCode=1 Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.340014 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.340090 4710 scope.go:117] "RemoveContainer" containerID="a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.341056 4710 scope.go:117] "RemoveContainer" containerID="e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.341262 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: E1002 12:50:13.341368 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.358366 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.373670 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.391554 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.407050 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.415986 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.416175 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.416259 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.416336 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.416411 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.424027 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.446715 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.466119 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a90cbe7d0994fac3fce03c07d2f10a03caab0646a3952dfe645f1e03e3a7f0b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:49:52Z\\\",\\\"message\\\":\\\"bad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 12:49:51.989327 6203 services_controller.go:360] Finished syncing service route-controller-manager on namespace openshift-route-controller-manager for network=default : 1.373434ms\\\\nI1002 12:49:51.990648 6203 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1002 12:49:51.990674 6203 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1002 12:49:51.990687 6203 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1002 12:49:51.990723 6203 factory.go:1336] Added *v1.Node event handler 7\\\\nI1002 12:49:51.990774 6203 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1002 12:49:51.991005 6203 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 12:49:51.991100 6203 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 12:49:51.991136 6203 ovnkube.go:599] Stopped ovnkube\\\\nI1002 12:49:51.991153 6203 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 12:49:51.991212 6203 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.477127 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.488922 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.501322 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.512688 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.519137 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.519163 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.519173 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.519196 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.519209 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.523229 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.534400 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.545209 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.556843 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.566580 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.577219 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.586403 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.596869 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.608319 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.617827 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.621246 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.621276 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.621287 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.621303 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.621313 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.723847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.724145 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.724251 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.724345 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.724445 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.826804 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.826847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.826856 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.826870 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.826880 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.903509 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:13 crc kubenswrapper[4710]: E1002 12:50:13.903711 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.929439 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.929651 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.929732 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.929909 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.929940 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.942472 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.942496 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.942506 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.942518 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.942528 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: E1002 12:50:13.956090 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.959786 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.959828 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.959843 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.959864 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.959881 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: E1002 12:50:13.972547 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.976281 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.976312 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.976323 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.976338 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.976350 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:13 crc kubenswrapper[4710]: E1002 12:50:13.991882 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:13Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.994836 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.994877 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.994886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.994902 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:13 crc kubenswrapper[4710]: I1002 12:50:13.994912 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:13Z","lastTransitionTime":"2025-10-02T12:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.008928 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.012810 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.012847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.012858 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.012873 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.012883 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.027039 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.027198 4710 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.032929 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.032978 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.032997 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.033020 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.033037 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.135818 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.135884 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.135899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.135920 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.135937 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.238289 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.238340 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.238353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.238370 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.238381 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.341042 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.341116 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.341131 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.341147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.341159 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.344608 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/2.log" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.347489 4710 scope.go:117] "RemoveContainer" containerID="e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a" Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.347654 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.360011 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.369433 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.379877 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.391774 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.414798 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.424906 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.435321 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.444164 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.444211 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.444227 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.444248 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.444263 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.449040 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.460346 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.470612 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.478755 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.491367 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.504695 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.524021 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.540800 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.546564 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.546608 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.546621 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.546640 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.546653 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.556961 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.571066 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:14Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.649027 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.649076 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.649089 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.649106 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.649116 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.751283 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.751321 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.751330 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.751346 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.751363 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.853504 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.853540 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.853553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.853570 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.853583 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.904210 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.904350 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.904554 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.904602 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.904621 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:14 crc kubenswrapper[4710]: E1002 12:50:14.904737 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.955703 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.955774 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.955790 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.955808 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:14 crc kubenswrapper[4710]: I1002 12:50:14.955821 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:14Z","lastTransitionTime":"2025-10-02T12:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.058947 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.058993 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.059007 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.059025 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.059036 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.161794 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.161842 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.161854 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.161869 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.161880 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.265097 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.265157 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.265171 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.265196 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.265209 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.368594 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.368681 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.368704 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.368740 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.368800 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.471695 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.471780 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.471800 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.471821 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.471835 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.574519 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.574561 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.574571 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.574588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.574599 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.677868 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.677930 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.677951 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.677976 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.677993 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.781352 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.781420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.781444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.781474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.781499 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.885039 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.885163 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.885185 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.885210 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.885260 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.904137 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:15 crc kubenswrapper[4710]: E1002 12:50:15.904378 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.988710 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.988826 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.988847 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.988877 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:15 crc kubenswrapper[4710]: I1002 12:50:15.988897 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:15Z","lastTransitionTime":"2025-10-02T12:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.091790 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.091878 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.091916 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.091949 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.091971 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.194840 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.194914 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.194938 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.194967 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.194989 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.298484 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.298565 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.298591 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.298622 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.298644 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.401860 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.401889 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.401900 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.401916 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.401928 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.504349 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.504385 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.504396 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.504411 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.504422 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.607532 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.607615 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.607637 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.607669 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.607692 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.711437 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.711509 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.711528 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.711557 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.711579 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.814310 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.814379 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.814397 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.814422 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.814444 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.903515 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.903534 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:16 crc kubenswrapper[4710]: E1002 12:50:16.903667 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.903776 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:16 crc kubenswrapper[4710]: E1002 12:50:16.903883 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:16 crc kubenswrapper[4710]: E1002 12:50:16.903956 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.917309 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.917420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.917442 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.917468 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.917485 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:16Z","lastTransitionTime":"2025-10-02T12:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.926731 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.946695 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.962272 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:16 crc kubenswrapper[4710]: I1002 12:50:16.985600 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.013193 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.019976 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.020024 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.020036 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.020054 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.020065 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.028315 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.042055 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.055584 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.068044 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.080632 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.092686 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.107572 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.120561 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.122317 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.122396 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.122409 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.122431 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.122463 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.134960 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.149164 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.164297 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.178992 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:17Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.225787 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.225841 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.225858 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.225884 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.225902 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.328260 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.328324 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.328396 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.328425 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.328448 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.431829 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.431876 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.431893 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.431922 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.431942 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.534742 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.534830 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.534848 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.534873 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.534893 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.636845 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.636891 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.636903 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.636920 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.636949 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.739824 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.739875 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.739886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.739902 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.739914 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.841995 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.842043 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.842054 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.842074 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.842085 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.903727 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:17 crc kubenswrapper[4710]: E1002 12:50:17.903901 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.945213 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.945666 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.945678 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.945696 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:17 crc kubenswrapper[4710]: I1002 12:50:17.945708 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:17Z","lastTransitionTime":"2025-10-02T12:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.049060 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.049221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.049249 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.049284 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.049306 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.152342 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.152378 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.152388 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.152404 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.152417 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.255900 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.255952 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.255964 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.255982 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.255993 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.358862 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.358941 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.358963 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.358992 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.359014 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.462103 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.462232 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.462259 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.462292 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.462316 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.565282 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.565349 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.565363 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.565382 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.565393 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.668657 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.668727 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.668737 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.668770 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.668780 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.783537 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.783585 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.783593 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.783608 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.783619 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.886700 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.886757 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.886768 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.886785 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.886794 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.904180 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.904331 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:18 crc kubenswrapper[4710]: E1002 12:50:18.904460 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.904503 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:18 crc kubenswrapper[4710]: E1002 12:50:18.904660 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:18 crc kubenswrapper[4710]: E1002 12:50:18.904733 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.990519 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.990638 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.990675 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.990716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:18 crc kubenswrapper[4710]: I1002 12:50:18.990791 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:18Z","lastTransitionTime":"2025-10-02T12:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.095059 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.095114 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.095134 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.095193 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.095217 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.198342 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.198394 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.198409 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.198430 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.198445 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.301565 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.301622 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.301638 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.301661 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.301677 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.404354 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.404423 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.404441 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.404473 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.404496 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.507178 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.507270 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.507291 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.507329 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.507355 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.611198 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.611248 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.611259 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.611278 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.611291 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.713642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.713692 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.713714 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.713733 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.713773 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.817006 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.817071 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.817095 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.817132 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.817148 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.903729 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:19 crc kubenswrapper[4710]: E1002 12:50:19.904508 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.920147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.920201 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.920215 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.920235 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:19 crc kubenswrapper[4710]: I1002 12:50:19.920250 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:19Z","lastTransitionTime":"2025-10-02T12:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.022740 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.022814 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.022827 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.022843 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.022855 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.129304 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.129365 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.129390 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.129412 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.129427 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.232052 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.232091 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.232103 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.232118 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.232128 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.335054 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.335094 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.335103 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.335120 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.335131 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.437327 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.437413 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.437438 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.437474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.437493 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.539507 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.539541 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.539551 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.539563 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.539572 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.642619 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.643024 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.643161 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.643326 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.643464 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.746735 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.746834 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.746851 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.746874 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.746899 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.850140 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.850539 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.850687 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.850933 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.851090 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.904218 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.904289 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:20 crc kubenswrapper[4710]: E1002 12:50:20.904378 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.904304 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:20 crc kubenswrapper[4710]: E1002 12:50:20.904470 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:20 crc kubenswrapper[4710]: E1002 12:50:20.905164 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.954650 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.954716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.954733 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.954792 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:20 crc kubenswrapper[4710]: I1002 12:50:20.954811 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:20Z","lastTransitionTime":"2025-10-02T12:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.058174 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.058338 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.058364 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.058448 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.058476 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.161902 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.161966 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.162021 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.162077 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.162094 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.267537 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.267624 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.267648 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.267677 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.267698 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.370503 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.370538 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.370546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.370558 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.370566 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.472712 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.472771 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.472784 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.472801 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.472814 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.575911 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.575962 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.575981 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.576004 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.576021 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.678833 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.678870 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.678882 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.678899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.678910 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.781352 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.781380 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.781387 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.781402 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.781411 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.883836 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.883895 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.883910 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.883932 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.883952 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.904380 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:21 crc kubenswrapper[4710]: E1002 12:50:21.904545 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.986729 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.986820 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.986836 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.986860 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:21 crc kubenswrapper[4710]: I1002 12:50:21.986876 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:21Z","lastTransitionTime":"2025-10-02T12:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.088463 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.088526 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.088550 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.088579 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.088602 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.191852 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.191961 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.191991 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.192026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.192049 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.297662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.297710 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.297733 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.297812 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.297833 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.399477 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.399518 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.399530 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.399546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.399557 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.502059 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.502444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.502463 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.502492 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.502514 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.605274 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.605323 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.605343 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.605368 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.605385 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.708410 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.708496 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.708515 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.708549 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.708571 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.811051 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.811127 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.811147 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.811181 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.811205 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.903803 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.903857 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.903904 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:22 crc kubenswrapper[4710]: E1002 12:50:22.904022 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:22 crc kubenswrapper[4710]: E1002 12:50:22.904142 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:22 crc kubenswrapper[4710]: E1002 12:50:22.904342 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.913113 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.913166 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.913178 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.913195 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:22 crc kubenswrapper[4710]: I1002 12:50:22.913206 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:22Z","lastTransitionTime":"2025-10-02T12:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.016798 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.016870 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.016882 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.016899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.016910 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.120122 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.120170 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.120184 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.120207 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.120224 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.223162 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.223229 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.223241 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.223265 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.223280 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.326550 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.326610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.326629 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.326656 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.326687 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.429928 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.429982 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.429997 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.430018 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.430028 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.533276 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.533639 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.534107 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.534244 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.534359 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.637184 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.637345 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.637407 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.637516 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.637588 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.740886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.740945 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.740955 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.740971 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.740981 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.843314 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.843353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.843361 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.843373 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.843382 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.873800 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:23 crc kubenswrapper[4710]: E1002 12:50:23.873965 4710 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:50:23 crc kubenswrapper[4710]: E1002 12:50:23.874034 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs podName:d77362b1-cb9c-498f-8e21-9114ce2de0d3 nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.874015059 +0000 UTC m=+99.980425942 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs") pod "network-metrics-daemon-zqx98" (UID: "d77362b1-cb9c-498f-8e21-9114ce2de0d3") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.903696 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:23 crc kubenswrapper[4710]: E1002 12:50:23.903860 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.945543 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.945580 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.945588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.945604 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:23 crc kubenswrapper[4710]: I1002 12:50:23.945613 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:23Z","lastTransitionTime":"2025-10-02T12:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.049621 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.050552 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.050869 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.051058 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.051264 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.155933 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.155995 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.156006 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.156024 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.156035 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.258725 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.258823 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.258836 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.258890 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.258905 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.351635 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.351697 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.351737 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.351790 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.351812 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.372202 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:24Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.376380 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.376426 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.376444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.376466 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.376483 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.392014 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:24Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.396930 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.397003 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.397028 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.397061 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.397088 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.417132 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:24Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.422205 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.422266 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.422286 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.422310 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.422330 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.442052 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:24Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.446383 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.446428 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.446442 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.446461 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.446473 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.468342 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:24Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.468551 4710 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.470026 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.470081 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.470097 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.470116 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.470131 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.572582 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.572632 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.572645 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.572659 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.572668 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.675660 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.675710 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.675723 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.675758 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.675770 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.779008 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.779067 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.779077 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.779094 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.779105 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.881785 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.881839 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.881849 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.881868 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.881881 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.903674 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.903776 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.903819 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.903957 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.904069 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:24 crc kubenswrapper[4710]: E1002 12:50:24.904185 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.984274 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.984313 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.984324 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.984350 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:24 crc kubenswrapper[4710]: I1002 12:50:24.984363 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:24Z","lastTransitionTime":"2025-10-02T12:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.085905 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.085943 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.085954 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.085969 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.085981 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.188170 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.188203 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.188211 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.188226 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.188237 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.291189 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.291241 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.291255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.291282 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.291298 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.393816 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.393886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.393905 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.393929 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.393947 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.497120 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.497219 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.497234 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.497260 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.497277 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.599252 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.599323 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.599356 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.599384 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.599402 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.702216 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.702258 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.702269 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.702286 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.702300 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.804607 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.804660 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.804672 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.804692 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.804705 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.904622 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:25 crc kubenswrapper[4710]: E1002 12:50:25.905183 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.907259 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.907319 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.907387 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.907416 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:25 crc kubenswrapper[4710]: I1002 12:50:25.907501 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:25Z","lastTransitionTime":"2025-10-02T12:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.010074 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.010144 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.010157 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.010184 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.010201 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.112362 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.112407 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.112419 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.112438 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.112451 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.214888 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.214959 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.214982 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.215012 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.215034 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.317967 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.318030 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.318054 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.318081 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.318100 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.388411 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8ddjb_7786a93b-a392-425e-b4ac-ed4956546d08/kube-multus/0.log" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.388471 4710 generic.go:334] "Generic (PLEG): container finished" podID="7786a93b-a392-425e-b4ac-ed4956546d08" containerID="b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d" exitCode=1 Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.388505 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8ddjb" event={"ID":"7786a93b-a392-425e-b4ac-ed4956546d08","Type":"ContainerDied","Data":"b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.388937 4710 scope.go:117] "RemoveContainer" containerID="b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.403890 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.417182 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.420333 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.420359 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.420370 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.420386 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.420398 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.429825 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.445836 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.459694 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.474191 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:25Z\\\",\\\"message\\\":\\\"2025-10-02T12:49:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4\\\\n2025-10-02T12:49:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4 to /host/opt/cni/bin/\\\\n2025-10-02T12:49:40Z [verbose] multus-daemon started\\\\n2025-10-02T12:49:40Z [verbose] Readiness Indicator file check\\\\n2025-10-02T12:50:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.484426 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.500592 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.515599 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.523677 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.523718 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.523727 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.523757 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.523766 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.532794 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.545788 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.561612 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.577448 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.587335 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.600527 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.615856 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.641190 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.641255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.641275 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.641303 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.641326 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.673986 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.743953 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.744001 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.744017 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.744044 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.744062 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.846012 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.846169 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.846229 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.846319 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.846412 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.903862 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:26 crc kubenswrapper[4710]: E1002 12:50:26.903975 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.904028 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.904035 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:26 crc kubenswrapper[4710]: E1002 12:50:26.904227 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:26 crc kubenswrapper[4710]: E1002 12:50:26.904262 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.915986 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.929713 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.942867 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.948667 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.948706 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.948729 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.948765 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.948779 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:26Z","lastTransitionTime":"2025-10-02T12:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.956211 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.968246 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.982655 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:26 crc kubenswrapper[4710]: I1002 12:50:26.993657 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.010241 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.032226 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.050570 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.050599 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.050608 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.050620 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.050630 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.060454 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.072691 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.085174 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.097874 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.111806 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:25Z\\\",\\\"message\\\":\\\"2025-10-02T12:49:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4\\\\n2025-10-02T12:49:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4 to /host/opt/cni/bin/\\\\n2025-10-02T12:49:40Z [verbose] multus-daemon started\\\\n2025-10-02T12:49:40Z [verbose] Readiness Indicator file check\\\\n2025-10-02T12:50:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.121873 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.133512 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.144106 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.153211 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.153271 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.153293 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.153320 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.153339 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.256596 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.256637 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.256649 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.256665 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.256677 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.361705 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.361968 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.362038 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.362108 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.362170 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.394106 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8ddjb_7786a93b-a392-425e-b4ac-ed4956546d08/kube-multus/0.log" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.394171 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8ddjb" event={"ID":"7786a93b-a392-425e-b4ac-ed4956546d08","Type":"ContainerStarted","Data":"581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.414693 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.433129 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.458149 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.471546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.471759 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.471851 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.471941 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.472024 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.476240 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.488698 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.502887 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.519561 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.538955 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.558272 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:25Z\\\",\\\"message\\\":\\\"2025-10-02T12:49:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4\\\\n2025-10-02T12:49:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4 to /host/opt/cni/bin/\\\\n2025-10-02T12:49:40Z [verbose] multus-daemon started\\\\n2025-10-02T12:49:40Z [verbose] Readiness Indicator file check\\\\n2025-10-02T12:50:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.573107 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.575489 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.575553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.575576 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.575611 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.575633 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.593825 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.610651 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.631912 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.650193 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.671802 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.678435 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.678526 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.678545 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.678575 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.678593 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.694533 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.708742 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:27Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.781604 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.781659 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.781679 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.781704 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.781723 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.884860 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.884904 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.884913 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.884926 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.884938 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.903838 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:27 crc kubenswrapper[4710]: E1002 12:50:27.903980 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.987257 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.987306 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.987331 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.987356 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:27 crc kubenswrapper[4710]: I1002 12:50:27.987372 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:27Z","lastTransitionTime":"2025-10-02T12:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.089081 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.089144 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.089160 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.089182 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.089198 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.192224 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.192298 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.192318 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.192341 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.192360 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.294662 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.294708 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.294720 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.294738 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.294772 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.396643 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.396711 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.396724 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.396767 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.396814 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.500022 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.500128 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.500141 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.500162 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.500183 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.603281 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.603312 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.603319 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.603367 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.603381 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.705841 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.705880 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.705890 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.705903 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.705912 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.813494 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.813558 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.813568 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.813585 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.813615 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.904116 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.904161 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.904116 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:28 crc kubenswrapper[4710]: E1002 12:50:28.904335 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:28 crc kubenswrapper[4710]: E1002 12:50:28.904411 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:28 crc kubenswrapper[4710]: E1002 12:50:28.904852 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.905497 4710 scope.go:117] "RemoveContainer" containerID="e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a" Oct 02 12:50:28 crc kubenswrapper[4710]: E1002 12:50:28.905945 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.915956 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.916185 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.916342 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.916484 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:28 crc kubenswrapper[4710]: I1002 12:50:28.916613 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:28Z","lastTransitionTime":"2025-10-02T12:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.019083 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.019111 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.019122 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.019136 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.019147 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.121726 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.121775 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.121785 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.121800 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.121812 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.223670 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.223712 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.223722 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.223737 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.223771 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.326953 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.327031 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.327055 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.327089 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.327113 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.430728 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.431087 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.431168 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.431248 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.431336 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.535449 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.536462 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.536511 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.536560 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.536577 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.639414 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.639508 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.639536 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.639573 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.639597 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.742861 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.742950 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.742973 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.743009 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.743033 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.846230 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.846591 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.846716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.846810 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.846886 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.903581 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:29 crc kubenswrapper[4710]: E1002 12:50:29.903726 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.949242 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.949272 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.949280 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.949295 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:29 crc kubenswrapper[4710]: I1002 12:50:29.949305 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:29Z","lastTransitionTime":"2025-10-02T12:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.051792 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.051839 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.051855 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.051871 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.051882 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.154142 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.154176 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.154185 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.154198 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.154207 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.256245 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.256331 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.256353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.256383 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.256400 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.359074 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.359127 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.359143 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.359163 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.359178 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.461349 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.461388 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.461398 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.461413 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.461425 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.564199 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.564259 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.564276 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.564299 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.564318 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.667080 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.667116 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.667127 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.667142 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.667154 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.769975 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.770041 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.770053 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.770088 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.770101 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.873048 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.873137 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.873161 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.873189 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.873215 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.904363 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:30 crc kubenswrapper[4710]: E1002 12:50:30.904576 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.904610 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:30 crc kubenswrapper[4710]: E1002 12:50:30.904730 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.904872 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:30 crc kubenswrapper[4710]: E1002 12:50:30.904974 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.977828 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.978275 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.978468 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.978665 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:30 crc kubenswrapper[4710]: I1002 12:50:30.978911 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:30Z","lastTransitionTime":"2025-10-02T12:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.082004 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.082040 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.082048 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.082064 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.082072 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.184344 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.184555 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.184642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.184739 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.184881 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.287489 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.287524 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.287534 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.287546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.287555 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.389917 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.389977 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.389993 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.390014 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.390030 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.492588 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.492635 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.492675 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.492697 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.492713 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.595504 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.595562 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.595579 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.595604 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.595621 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.697906 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.697980 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.698000 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.698020 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.698034 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.800739 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.800880 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.800899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.800923 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.800940 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.903347 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:31 crc kubenswrapper[4710]: E1002 12:50:31.903474 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.903790 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.903823 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.903832 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.903843 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:31 crc kubenswrapper[4710]: I1002 12:50:31.903853 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:31Z","lastTransitionTime":"2025-10-02T12:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.006988 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.007047 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.007058 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.007081 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.007093 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.109307 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.109370 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.109393 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.109421 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.109442 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.212716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.212761 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.212769 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.212782 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.212791 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.315503 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.315538 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.315546 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.315560 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.315572 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.418510 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.418566 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.418582 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.418606 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.418625 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.521187 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.521244 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.521262 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.521287 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.521307 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.623906 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.623972 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.623989 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.624014 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.624031 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.726396 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.726444 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.726457 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.726474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.726489 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.828452 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.828492 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.828505 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.828534 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.828555 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.904269 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.904328 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.904345 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:32 crc kubenswrapper[4710]: E1002 12:50:32.904405 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:32 crc kubenswrapper[4710]: E1002 12:50:32.904468 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:32 crc kubenswrapper[4710]: E1002 12:50:32.904591 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.931244 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.931289 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.931299 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.931316 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:32 crc kubenswrapper[4710]: I1002 12:50:32.931328 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:32Z","lastTransitionTime":"2025-10-02T12:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.034392 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.034445 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.034457 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.034478 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.034493 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.136924 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.136983 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.136996 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.137018 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.137032 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.239636 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.239671 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.239683 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.239698 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.239709 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.342463 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.342505 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.342514 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.342529 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.342540 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.446511 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.446562 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.446574 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.446590 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.446600 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.550025 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.550087 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.550104 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.550130 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.550151 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.653439 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.653483 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.653495 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.653512 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.653526 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.755628 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.755661 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.755672 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.755688 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.755699 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.858862 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.858902 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.858913 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.858928 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.858940 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.903673 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:33 crc kubenswrapper[4710]: E1002 12:50:33.903947 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.961415 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.961466 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.961487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.961511 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:33 crc kubenswrapper[4710]: I1002 12:50:33.961531 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:33Z","lastTransitionTime":"2025-10-02T12:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.064808 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.064844 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.064857 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.064873 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.064884 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.167493 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.167554 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.167576 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.167604 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.167624 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.271046 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.271108 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.271122 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.271140 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.271154 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.374915 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.375439 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.375631 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.375825 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.375980 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.479634 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.479677 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.479690 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.479707 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.479720 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.506829 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.506899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.506919 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.506946 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.506964 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.528295 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:34Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.533146 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.533221 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.533245 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.533276 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.533296 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.554472 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:34Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.559984 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.560052 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.560078 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.560105 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.560127 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.579441 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:34Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.584489 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.584543 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.584560 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.584585 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.584605 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.601885 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:34Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.608192 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.608248 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.608267 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.608291 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.608309 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.628534 4710 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"da9dd993-7c58-4d65-9aab-f7f0f0784918\\\",\\\"systemUUID\\\":\\\"5d561d62-be46-4109-8ed7-b35f0572f5b0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:34Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.628816 4710 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.631371 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.631424 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.631445 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.631471 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.631489 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.734542 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.734603 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.734624 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.734653 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.734682 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.837539 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.837600 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.837622 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.837652 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.837680 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.903846 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.903852 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.903881 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.904194 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.904332 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:34 crc kubenswrapper[4710]: E1002 12:50:34.904558 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.940377 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.940432 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.940449 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.940474 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:34 crc kubenswrapper[4710]: I1002 12:50:34.940492 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:34Z","lastTransitionTime":"2025-10-02T12:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.043914 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.043974 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.043995 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.044025 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.044047 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.147120 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.147181 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.147199 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.147224 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.147245 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.250123 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.250179 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.250194 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.250218 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.250234 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.353338 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.353389 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.353399 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.353416 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.353428 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.455633 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.455667 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.455679 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.455695 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.455706 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.557459 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.557518 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.557538 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.557556 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.557568 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.660041 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.660112 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.660127 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.660146 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.660160 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.762005 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.762042 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.762051 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.762064 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.762072 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.864879 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.864919 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.864927 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.864958 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.864968 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.903987 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:35 crc kubenswrapper[4710]: E1002 12:50:35.904223 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.967499 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.967550 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.967607 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.967650 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:35 crc kubenswrapper[4710]: I1002 12:50:35.967675 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:35Z","lastTransitionTime":"2025-10-02T12:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.070819 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.070913 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.070934 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.070959 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.071012 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.174417 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.174470 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.174487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.174509 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.174525 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.277826 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.277900 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.277927 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.278085 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.278128 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.381820 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.381899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.381922 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.381951 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.381972 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.484666 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.484749 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.484806 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.484829 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.484849 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.588458 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.588576 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.588598 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.588621 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.588639 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.692937 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.693021 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.693043 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.693072 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.693096 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.796691 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.796796 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.796822 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.796853 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.796877 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.899538 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.899590 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.899601 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.899618 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.899645 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:36Z","lastTransitionTime":"2025-10-02T12:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.903732 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.903773 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:36 crc kubenswrapper[4710]: E1002 12:50:36.903845 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:36 crc kubenswrapper[4710]: E1002 12:50:36.904146 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.904453 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:36 crc kubenswrapper[4710]: E1002 12:50:36.904646 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.918047 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:36Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.937393 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eda4d6c2-f0cd-43e6-b16e-44c5147e5281\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e80f000622d2bd6d53a612d602cfa68bbaadccbc49041633902ec5cacfee8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589d3c13430b2361ccce91b99512d4bf9b00909c8e983a69df923cab8ea1d4fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3f30809c2bc73beb61a57fc76aa5346047b5dc44283ef04d27ebae88797053\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ff8b42373f12ac27a3ff486e335f9840d9338d0ad0a7b2825c9666b699aa1e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3909e02501686d662fd034af4404e54985d05305b7464b014295ef2170ccd814\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbf3e40128fbb0c375e3a6053c5a8f61e8daf6cf3b51400bc3664aa2c531474\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a3008343e4b28b5873bb35f4eb72a042796d8d84543a3d2e3b47951ec787cf8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhlsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qpqlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:36Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.957018 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:12Z\\\",\\\"message\\\":\\\"t handler 1 for removal\\\\nI1002 12:50:12.929463 6412 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 12:50:12.929489 6412 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 12:50:12.929581 6412 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 12:50:12.929600 6412 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 12:50:12.929602 6412 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 12:50:12.929609 6412 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 12:50:12.929617 6412 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 12:50:12.929775 6412 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 12:50:12.929939 6412 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.929979 6412 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930128 6412 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 12:50:12.930507 6412 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8jhhq_openshift-ovn-kubernetes(93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wlht4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8jhhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:36Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.969775 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8ddjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7786a93b-a392-425e-b4ac-ed4956546d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T12:50:25Z\\\",\\\"message\\\":\\\"2025-10-02T12:49:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4\\\\n2025-10-02T12:49:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_126b35ca-cd5d-45b6-a095-2ae34fe974c4 to /host/opt/cni/bin/\\\\n2025-10-02T12:49:40Z [verbose] multus-daemon started\\\\n2025-10-02T12:49:40Z [verbose] Readiness Indicator file check\\\\n2025-10-02T12:50:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-km2kc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8ddjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:36Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.980000 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zmkll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d709a39d-880c-4499-8429-cd1c815af683\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72eb4cd787ef4d62bba42ac6fa89f68e65e510331ae9352f444875374b5405d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zmkll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:36Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:36 crc kubenswrapper[4710]: I1002 12:50:36.990619 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aed137d-9ef0-4e9b-b9d2-b6dca3e2defe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f297e38c24917f2f0a74da7d63617e6fd7495ca3a13de094454600d406b3127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb9876c747aa8143a3fe58ec08553d22ead5143c34fb78a6a6260382f23edf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plq59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nwkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:36Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.001509 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.001556 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.001572 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.001593 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.001608 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.004038 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zqx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77362b1-cb9c-498f-8e21-9114ce2de0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhd24\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zqx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.023295 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fb741ca-9c37-4fbf-b9e3-d88abbbe1430\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd58fd81c7cd7b43d7f34b848073390b542cf90c18e16cdc92952c105e8fbc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f8e290fca70d1c503f3df38d29a503ee2e12c91b39876464fc07b3b4f31ab6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be9cc0a6d1c9733953a552c5a482367d9f901fed7931a18d1aca0617f2380d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75dd3780d05772e2d0bcd4ebf03826d16e7536b1cf879ada0f8a84ac38db20fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.039192 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47f1f670a89f3ebfa5dec74afe0ca0f55fd601c06006a2c7a9eb701f075e42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://097c897d6b52065e8ce77b184d2a3791d3deb24ec9f1980492456ce334a40055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.058963 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.073930 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96855e142e88ced841de94251eaef25bfde2f178efc5041af5e5883374644f93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.091605 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.105235 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.105293 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.105313 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.105336 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.105353 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.109929 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01947451-6af9-4b65-b74c-87e098437b71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c255094d56e71805059acad6343d4d0574b460bdbf60f5e4ad3a9db21e09563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6fbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tblbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.128313 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5f25b29-1826-45a0-ad62-044bc03277f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c49ac3453c2c65e110ce9548788912ffd09bb19b0488ed2624b89807eb69ea09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51d4b42702c5fe35f271a8199d27e971a039eda7f9c2d68725b69c31dbebb06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c49a302d86df4b45401d63130aaac867431e86b1d160a363c253de87d5ac30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb92c91c57fa4ab20ab8d0d6b3367f30b68c3bed59635feafc64193664f4b8e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268552d951e2dda2d872c5d563b0c20f3042cb7cf515fb3729dff778ee65cabd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T12:49:36Z\\\",\\\"message\\\":\\\"t denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 12:49:36.875142 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 12:49:36.875170 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875175 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 12:49:36.875179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 12:49:36.875182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 12:49:36.875184 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 12:49:36.875188 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 12:49:36.875456 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 12:49:36.881456 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1556666359/tls.crt::/tmp/serving-cert-1556666359/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759409360\\\\\\\\\\\\\\\" (2025-10-02 12:49:19 +0000 UTC to 2025-11-01 12:49:20 +0000 UTC (now=2025-10-02 12:49:36.881421439 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881617 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759409371\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759409370\\\\\\\\\\\\\\\" (2025-10-02 11:49:30 +0000 UTC to 2026-10-02 11:49:30 +0000 UTC (now=2025-10-02 12:49:36.881599373 +0000 UTC))\\\\\\\"\\\\nI1002 12:49:36.881669 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1002 12:49:36.881711 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1002 12:49:36.881848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9898bd46bd4179283f651d3e2a8718bdd33d7123928d99903eccb2f486cc589\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2feead31cd1ee555454faf406f60350deef9e68f1c170439ec7671727689c909\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.146049 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7296f149-fae9-4f7c-be49-9279a353e4da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec13164a7dfeeb3e3f41191c0fe0a9a64bf4da5d85a6aa82c48dd7806d81cc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23af2f220444b10cc5e17a400872677b0f8f0d965a163bd1c03f195b47744f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dafd7f08ff0c8020937fc11d3216f637d3136b210b9423aded2d61c33b08eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c5f973ea9af4a16578b668d6b2d5e408e2ed8f32f553a2da19787eae2c87664\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T12:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T12:49:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.163410 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d60025ec95037e6e653ce4a5e64497d28428d8d29fdc348a68e6e13635f2362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.179695 4710 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-skmsh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74797a-11a0-4174-b10d-0c6a63566f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T12:49:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a24b5d4dc4acf087f41d452a9f2ae0583e91b8e6d0c9ff7335610d1228d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T12:49:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T12:49:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-skmsh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T12:50:37Z is after 2025-08-24T17:21:41Z" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.207721 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.207773 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.207784 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.207798 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.207809 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.311102 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.311160 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.311180 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.311206 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.311227 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.414645 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.415159 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.415176 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.415198 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.415214 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.518424 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.518470 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.518479 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.518493 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.518503 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.621705 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.621813 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.621838 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.621865 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.621887 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.724656 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.724720 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.724736 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.724791 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.724810 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.828348 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.828500 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.828520 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.828553 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.828572 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.904354 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:37 crc kubenswrapper[4710]: E1002 12:50:37.904647 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.932163 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.932244 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.932280 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.932311 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:37 crc kubenswrapper[4710]: I1002 12:50:37.932332 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:37Z","lastTransitionTime":"2025-10-02T12:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.035283 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.035342 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.035359 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.035383 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.035401 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.138413 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.138478 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.138495 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.138521 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.138539 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.242098 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.242191 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.242218 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.242255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.242280 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.345330 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.345378 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.345390 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.345409 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.345420 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.447882 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.447952 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.447971 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.447999 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.448018 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.551671 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.551729 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.551779 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.551806 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.551823 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.654890 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.654957 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.654980 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.655008 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.655030 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.758306 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.758351 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.758360 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.758378 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.758389 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.861395 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.861446 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.861465 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.861490 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.861508 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.904603 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.904742 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:38 crc kubenswrapper[4710]: E1002 12:50:38.904989 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.905051 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:38 crc kubenswrapper[4710]: E1002 12:50:38.905231 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:38 crc kubenswrapper[4710]: E1002 12:50:38.905287 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.965112 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.965161 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.965176 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.965196 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:38 crc kubenswrapper[4710]: I1002 12:50:38.965211 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:38Z","lastTransitionTime":"2025-10-02T12:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.067858 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.067895 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.067904 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.067919 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.067927 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.171370 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.171424 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.171442 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.171469 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.171488 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.274623 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.275138 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.275302 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.275627 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.275813 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.378801 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.379082 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.379180 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.379276 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.379346 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.482255 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.482361 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.482384 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.482415 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.482438 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.586053 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.586106 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.586118 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.586137 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.586149 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.688730 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.688821 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.688837 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.688859 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.688875 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.792552 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.792600 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.792612 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.792634 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.792653 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.896523 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.896624 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.896668 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.896702 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.896720 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:39Z","lastTransitionTime":"2025-10-02T12:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:39 crc kubenswrapper[4710]: I1002 12:50:39.903930 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:39 crc kubenswrapper[4710]: E1002 12:50:39.904297 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.000618 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.000664 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.000678 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.000699 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.000710 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.103993 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.104044 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.104060 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.104081 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.104093 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.207029 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.207110 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.207133 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.207162 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.207179 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.310381 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.310450 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.310468 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.310491 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.310509 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.413727 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.413828 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.413853 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.413884 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.413908 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.516865 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.516913 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.516925 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.516941 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.516952 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.620332 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.620397 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.620414 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.620439 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.620458 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.723814 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.723870 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.723886 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.723909 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.723926 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.736392 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.736630 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:51:44.736593806 +0000 UTC m=+148.843004729 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.827181 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.827238 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.827271 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.827299 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.827321 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.837785 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.837827 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.837851 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.837886 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.837913 4710 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838028 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:51:44.837997202 +0000 UTC m=+148.944408115 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838141 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838163 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838173 4710 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838205 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838244 4710 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838264 4710 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838269 4710 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838223 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 12:51:44.838209407 +0000 UTC m=+148.944620290 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838312 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 12:51:44.838306179 +0000 UTC m=+148.944717062 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.838381 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 12:51:44.838355091 +0000 UTC m=+148.944766054 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.905875 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.906053 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.906314 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.906669 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.906930 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:40 crc kubenswrapper[4710]: E1002 12:50:40.907463 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.909261 4710 scope.go:117] "RemoveContainer" containerID="e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.928428 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.929835 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.930043 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.930190 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.930342 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:40 crc kubenswrapper[4710]: I1002 12:50:40.930489 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:40Z","lastTransitionTime":"2025-10-02T12:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.033053 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.033487 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.033543 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.033566 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.033613 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.137214 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.137293 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.137313 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.137351 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.137369 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.241034 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.241095 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.241111 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.241137 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.241156 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.344526 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.344575 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.344587 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.344610 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.344627 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.446526 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.446589 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.446608 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.446631 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.446647 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.449828 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/2.log" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.452985 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerStarted","Data":"0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.453953 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.501812 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=64.501790397 podStartE2EDuration="1m4.501790397s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.48915431 +0000 UTC m=+85.595565213" watchObservedRunningTime="2025-10-02 12:50:41.501790397 +0000 UTC m=+85.608201280" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.502059 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=28.502055364 podStartE2EDuration="28.502055364s" podCreationTimestamp="2025-10-02 12:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.501983202 +0000 UTC m=+85.608394115" watchObservedRunningTime="2025-10-02 12:50:41.502055364 +0000 UTC m=+85.608466247" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.549220 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.549280 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.549300 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.549326 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.549340 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.581377 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podStartSLOduration=64.581361295 podStartE2EDuration="1m4.581361295s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.564523353 +0000 UTC m=+85.670934236" watchObservedRunningTime="2025-10-02 12:50:41.581361295 +0000 UTC m=+85.687772168" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.594970 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-skmsh" podStartSLOduration=64.594950036 podStartE2EDuration="1m4.594950036s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.594467714 +0000 UTC m=+85.700878597" watchObservedRunningTime="2025-10-02 12:50:41.594950036 +0000 UTC m=+85.701360929" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.650997 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qpqlq" podStartSLOduration=64.650971903 podStartE2EDuration="1m4.650971903s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.626985871 +0000 UTC m=+85.733396754" watchObservedRunningTime="2025-10-02 12:50:41.650971903 +0000 UTC m=+85.757382786" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.651814 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.651876 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.651894 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.651922 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.651940 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.670809 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podStartSLOduration=64.670786181 podStartE2EDuration="1m4.670786181s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.651579138 +0000 UTC m=+85.757990031" watchObservedRunningTime="2025-10-02 12:50:41.670786181 +0000 UTC m=+85.777197064" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.671200 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.671196121 podStartE2EDuration="1.671196121s" podCreationTimestamp="2025-10-02 12:50:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.670423411 +0000 UTC m=+85.776834294" watchObservedRunningTime="2025-10-02 12:50:41.671196121 +0000 UTC m=+85.777607004" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.706361 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.706346883 podStartE2EDuration="1m4.706346883s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.684331461 +0000 UTC m=+85.790742354" watchObservedRunningTime="2025-10-02 12:50:41.706346883 +0000 UTC m=+85.812757766" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.721675 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8ddjb" podStartSLOduration=64.721658468 podStartE2EDuration="1m4.721658468s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.720805746 +0000 UTC m=+85.827216639" watchObservedRunningTime="2025-10-02 12:50:41.721658468 +0000 UTC m=+85.828069341" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.736594 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zmkll" podStartSLOduration=64.736575272 podStartE2EDuration="1m4.736575272s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.736346327 +0000 UTC m=+85.842757230" watchObservedRunningTime="2025-10-02 12:50:41.736575272 +0000 UTC m=+85.842986155" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.754835 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.754872 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.754885 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.754902 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.754913 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.763863 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nwkn4" podStartSLOduration=63.763846307 podStartE2EDuration="1m3.763846307s" podCreationTimestamp="2025-10-02 12:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:41.75120506 +0000 UTC m=+85.857615943" watchObservedRunningTime="2025-10-02 12:50:41.763846307 +0000 UTC m=+85.870257200" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.817888 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zqx98"] Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.818019 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:41 crc kubenswrapper[4710]: E1002 12:50:41.818112 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.857785 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.857824 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.857836 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.857850 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.857860 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.961183 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.961664 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.961678 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.961701 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:41 crc kubenswrapper[4710]: I1002 12:50:41.961929 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:41Z","lastTransitionTime":"2025-10-02T12:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.064978 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.065028 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.065037 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.065053 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.065062 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.167799 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.167903 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.167934 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.167972 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.168001 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.270479 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.270525 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.270536 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.270550 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.270581 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.374067 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.374122 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.374133 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.374155 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.374168 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.476786 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.476840 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.476854 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.476874 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.476890 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.585420 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.585501 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.585518 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.585544 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.585563 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.689150 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.689231 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.689249 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.689275 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.689292 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.792811 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.792853 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.792863 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.792879 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.792890 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.896600 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.896665 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.896682 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.896707 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.896723 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.904418 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.904416 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.904416 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:42 crc kubenswrapper[4710]: E1002 12:50:42.904837 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 12:50:42 crc kubenswrapper[4710]: E1002 12:50:42.904972 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 12:50:42 crc kubenswrapper[4710]: E1002 12:50:42.905087 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.999599 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.999673 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.999691 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.999716 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:42 crc kubenswrapper[4710]: I1002 12:50:42.999735 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:42Z","lastTransitionTime":"2025-10-02T12:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.103281 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.103331 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.103347 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.103374 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.103395 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.206306 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.206479 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.206559 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.206640 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.206768 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.309736 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.309844 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.309866 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.309899 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.309925 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.413622 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.413722 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.413789 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.413824 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.413849 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.516729 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.516815 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.516832 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.516856 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.516873 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.619200 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.619282 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.619307 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.619329 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.619342 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.722665 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.722790 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.722817 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.722842 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.722859 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.826738 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.826880 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.826912 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.826941 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.826965 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.903539 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:43 crc kubenswrapper[4710]: E1002 12:50:43.903868 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zqx98" podUID="d77362b1-cb9c-498f-8e21-9114ce2de0d3" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.930267 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.930322 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.930335 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.930353 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:43 crc kubenswrapper[4710]: I1002 12:50:43.930366 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:43Z","lastTransitionTime":"2025-10-02T12:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.033574 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.033642 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.033658 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.033707 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.033728 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:44Z","lastTransitionTime":"2025-10-02T12:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.135951 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.136037 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.136059 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.136090 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.136111 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:44Z","lastTransitionTime":"2025-10-02T12:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.239732 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.239797 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.239807 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.239826 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.239835 4710 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T12:50:44Z","lastTransitionTime":"2025-10-02T12:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.343206 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.343273 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.343284 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.343304 4710 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.343465 4710 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.397832 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.398477 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.403217 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.411806 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.411866 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.413241 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gp652"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.413931 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.422488 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.423499 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.423566 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.423708 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.424633 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.424926 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.425247 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.425514 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.425829 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.426132 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mmjq2"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.426183 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.426445 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.426662 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-9cmjq"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.426738 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.427180 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.427201 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.427302 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.428582 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.432953 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-284s4"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.433563 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.436138 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.436416 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.436595 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.438797 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gqwlv"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.438976 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.439055 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.443460 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.443758 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.444107 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.444468 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.445014 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.450398 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.450536 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.451023 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.451769 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-m5hbd"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.451961 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.452233 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.452273 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.453187 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.453330 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cx6br"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.454294 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.479949 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.480439 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.481207 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c5sq4"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.481430 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.482100 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.482304 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.482436 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.483064 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.483872 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b7kkk"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.484277 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.484443 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.484785 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mrtmb"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.485248 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.485717 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.486269 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.487022 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.492836 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-46nzg"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.497698 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hq8zf"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.498133 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.498228 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-49mwb"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.500518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.501869 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.498322 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.502455 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.511835 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5mtxm"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.512730 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.516714 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.517498 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.519471 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.519626 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.519889 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520095 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520144 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520159 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520196 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520200 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520260 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520294 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520145 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520360 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520422 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520526 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.520856 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.521000 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.521194 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.521240 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.522390 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.522637 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-j786v"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.524334 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.524957 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.525627 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.529606 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.530538 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.532804 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.533344 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.534144 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.534320 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.534479 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.534645 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.534882 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.535244 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.537062 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.537415 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.547365 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.550189 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.551428 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.551700 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.563908 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.569829 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.570659 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.571076 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.571342 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.571560 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.571814 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.578200 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.578667 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.581170 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.581448 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.581611 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.581713 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.581951 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.582046 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.582134 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.582257 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.582422 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.582887 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583026 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583127 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583249 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583384 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583562 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583649 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583793 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.583992 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.585684 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-audit\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.585871 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdd48c4-7366-4e02-9cf9-78cb66e42249-serving-cert\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.585950 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntnwv\" (UniqueName: \"kubernetes.io/projected/d6d57748-0dc4-4ea8-9da5-2e6488933815-kube-api-access-ntnwv\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586035 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-config\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586099 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58f270a4-ab6a-4810-b99b-2af72f727a15-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586181 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90a304b8-7bb2-4d6b-885a-cea681b8d365-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586305 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90a304b8-7bb2-4d6b-885a-cea681b8d365-config\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586350 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7btxx\" (UniqueName: \"kubernetes.io/projected/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-kube-api-access-7btxx\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586383 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-etcd-client\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586630 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8q9p\" (UniqueName: \"kubernetes.io/projected/6cdd48c4-7366-4e02-9cf9-78cb66e42249-kube-api-access-q8q9p\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586654 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-image-import-ca\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586676 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-serving-cert\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586709 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586792 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2qz9\" (UniqueName: \"kubernetes.io/projected/58f270a4-ab6a-4810-b99b-2af72f727a15-kube-api-access-g2qz9\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586853 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6d57748-0dc4-4ea8-9da5-2e6488933815-trusted-ca\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586891 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2npwk\" (UniqueName: \"kubernetes.io/projected/90a304b8-7bb2-4d6b-885a-cea681b8d365-kube-api-access-2npwk\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586913 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58f270a4-ab6a-4810-b99b-2af72f727a15-metrics-tls\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586929 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/58f270a4-ab6a-4810-b99b-2af72f727a15-trusted-ca\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586952 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.586973 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d57748-0dc4-4ea8-9da5-2e6488933815-config\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587021 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwkg\" (UniqueName: \"kubernetes.io/projected/c289070e-b380-4303-ae1a-b21408ae9f32-kube-api-access-dwwkg\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587086 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-proxy-tls\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587108 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-encryption-config\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587117 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587140 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-client-ca\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587167 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-config\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587185 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c289070e-b380-4303-ae1a-b21408ae9f32-audit-dir\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587217 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587245 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6d57748-0dc4-4ea8-9da5-2e6488933815-serving-cert\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587269 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-etcd-serving-ca\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587289 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c289070e-b380-4303-ae1a-b21408ae9f32-node-pullsecrets\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587490 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587883 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587932 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.587977 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.588074 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.588221 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.588238 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.588281 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.588382 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.588408 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.591846 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.602034 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.605569 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.605835 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.606713 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cfvqq"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.607302 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.607311 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.607358 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.608031 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.608906 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.610097 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.610852 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.611100 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.611375 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.611549 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.613308 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.613966 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.617417 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.618890 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.619682 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.620212 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.620483 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.620516 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-c7t49"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.622805 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.623281 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.624450 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.624930 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.625678 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.630063 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.630119 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mmjq2"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.630233 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.631565 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.638575 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.645458 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.645773 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.649473 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.649495 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.649495 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.655180 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.656522 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-fq5v2"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.657431 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.657506 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-m5hbd"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.658357 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-284s4"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.659303 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-9cmjq"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.660092 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.662107 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vpc2k"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.663331 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.663678 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-49mwb"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.664295 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.665196 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-46nzg"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.666166 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.666986 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mrtmb"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.667917 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.668727 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.669628 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b7kkk"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.670537 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cx6br"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.671482 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.672418 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cfvqq"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.673329 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.674214 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.675078 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-xt66p"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.675898 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.675999 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hq8zf"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.676885 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.677796 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.678677 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.679575 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.680538 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gqwlv"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.681369 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.682262 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.683160 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.684213 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.684982 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.685892 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xt66p"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.686823 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c5sq4"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.687730 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688032 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-proxy-tls\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688061 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-encryption-config\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688080 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-client-ca\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688098 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c289070e-b380-4303-ae1a-b21408ae9f32-audit-dir\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688115 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-config\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688133 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688147 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-etcd-serving-ca\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688162 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6d57748-0dc4-4ea8-9da5-2e6488933815-serving-cert\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688178 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c289070e-b380-4303-ae1a-b21408ae9f32-node-pullsecrets\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688195 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-audit\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688210 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdd48c4-7366-4e02-9cf9-78cb66e42249-serving-cert\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688224 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntnwv\" (UniqueName: \"kubernetes.io/projected/d6d57748-0dc4-4ea8-9da5-2e6488933815-kube-api-access-ntnwv\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688242 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-config\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688259 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58f270a4-ab6a-4810-b99b-2af72f727a15-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688277 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90a304b8-7bb2-4d6b-885a-cea681b8d365-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688294 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90a304b8-7bb2-4d6b-885a-cea681b8d365-config\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688311 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7btxx\" (UniqueName: \"kubernetes.io/projected/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-kube-api-access-7btxx\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688327 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-etcd-client\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688344 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8q9p\" (UniqueName: \"kubernetes.io/projected/6cdd48c4-7366-4e02-9cf9-78cb66e42249-kube-api-access-q8q9p\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688361 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-image-import-ca\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688375 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-serving-cert\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688394 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688412 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2qz9\" (UniqueName: \"kubernetes.io/projected/58f270a4-ab6a-4810-b99b-2af72f727a15-kube-api-access-g2qz9\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688426 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6d57748-0dc4-4ea8-9da5-2e6488933815-trusted-ca\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688457 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/58f270a4-ab6a-4810-b99b-2af72f727a15-trusted-ca\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688468 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c289070e-b380-4303-ae1a-b21408ae9f32-node-pullsecrets\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688476 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2npwk\" (UniqueName: \"kubernetes.io/projected/90a304b8-7bb2-4d6b-885a-cea681b8d365-kube-api-access-2npwk\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688493 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58f270a4-ab6a-4810-b99b-2af72f727a15-metrics-tls\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688511 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688526 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d57748-0dc4-4ea8-9da5-2e6488933815-config\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688546 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwkg\" (UniqueName: \"kubernetes.io/projected/c289070e-b380-4303-ae1a-b21408ae9f32-kube-api-access-dwwkg\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.688776 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c289070e-b380-4303-ae1a-b21408ae9f32-audit-dir\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.689730 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-audit\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.689795 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-config\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.690053 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-etcd-serving-ca\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.690235 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-client-ca\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.690629 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5mtxm"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.690685 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xmcpk"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.691174 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90a304b8-7bb2-4d6b-885a-cea681b8d365-config\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.691453 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/58f270a4-ab6a-4810-b99b-2af72f727a15-trusted-ca\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.691529 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-c7t49"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.691616 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.692045 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.692561 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-config\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.693241 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.693436 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.693572 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-encryption-config\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.693889 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6d57748-0dc4-4ea8-9da5-2e6488933815-serving-cert\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694228 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d57748-0dc4-4ea8-9da5-2e6488933815-config\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694384 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694421 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fq5v2"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694642 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c289070e-b380-4303-ae1a-b21408ae9f32-image-import-ca\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694692 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-proxy-tls\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694970 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6d57748-0dc4-4ea8-9da5-2e6488933815-trusted-ca\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.694971 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90a304b8-7bb2-4d6b-885a-cea681b8d365-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.695126 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xmcpk"] Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.695156 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-serving-cert\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.695361 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdd48c4-7366-4e02-9cf9-78cb66e42249-serving-cert\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.696720 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c289070e-b380-4303-ae1a-b21408ae9f32-etcd-client\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.697574 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58f270a4-ab6a-4810-b99b-2af72f727a15-metrics-tls\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.700725 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.722129 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.740602 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.760612 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.780895 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.800317 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.821285 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.840095 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.860231 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.880674 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.900381 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.903598 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.903598 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.903777 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.920524 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.941277 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.960466 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 12:50:44 crc kubenswrapper[4710]: I1002 12:50:44.980558 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.000454 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.020693 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.041356 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.061566 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.081430 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.101194 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.121081 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.141552 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.161731 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.181276 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.201677 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.221087 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.242021 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.261430 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.281690 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.301194 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.320210 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.341132 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.360582 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.381502 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.421737 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.441632 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.460971 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.481106 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.502382 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.523178 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.541710 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.561390 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.581435 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.601069 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.619271 4710 request.go:700] Waited for 1.01003478s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/configmaps?fieldSelector=metadata.name%3Dmachine-config-operator-images&limit=500&resourceVersion=0 Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.621065 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.642561 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.661249 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.682376 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.715895 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.721283 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.742367 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.761491 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.781100 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.801278 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.820613 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.841649 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.862028 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.881414 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.901631 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.904072 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.921239 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.941135 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.962626 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 12:50:45 crc kubenswrapper[4710]: I1002 12:50:45.981288 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.001656 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.022065 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.041377 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.062392 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.081198 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.101459 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.121207 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.142152 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.161501 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.181344 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.202122 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.222228 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.241361 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.260596 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.281256 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.301567 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.320613 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.341440 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.361264 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.382010 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.401895 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.421785 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.441489 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.462377 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510534 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-tls\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510594 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-trusted-ca\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510632 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c876eadd-e408-42f5-9d74-bf570af2bfae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510673 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-signing-key\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510705 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/148b577d-5144-4ff2-bacf-a8a0fad52c61-images\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510799 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3b4d1d22-d5be-42d8-ad04-a8eda187570a-srv-cert\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510851 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510890 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsp64\" (UniqueName: \"kubernetes.io/projected/148b577d-5144-4ff2-bacf-a8a0fad52c61-kube-api-access-lsp64\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510923 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-encryption-config\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.510957 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72c8ef58-7bca-43c3-900d-297de2e48955-auth-proxy-config\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511009 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42nk\" (UniqueName: \"kubernetes.io/projected/e25fa232-dfb3-41d4-a36a-0d5723a67ca3-kube-api-access-x42nk\") pod \"multus-admission-controller-857f4d67dd-46nzg\" (UID: \"e25fa232-dfb3-41d4-a36a-0d5723a67ca3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511080 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88663a9d-6b82-4b98-a1a2-7629c72c28a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511127 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db597145-79db-4f18-ae0c-c64f35af2fcc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511159 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-oauth-serving-cert\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511191 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/148b577d-5144-4ff2-bacf-a8a0fad52c61-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511238 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511359 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: E1002 12:50:46.511424 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.011398652 +0000 UTC m=+91.117809575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511480 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brd76\" (UniqueName: \"kubernetes.io/projected/88663a9d-6b82-4b98-a1a2-7629c72c28a9-kube-api-access-brd76\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511549 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5px9\" (UniqueName: \"kubernetes.io/projected/c68e372e-5db7-475e-bacb-5452d32cbdfc-kube-api-access-p5px9\") pod \"downloads-7954f5f757-m5hbd\" (UID: \"c68e372e-5db7-475e-bacb-5452d32cbdfc\") " pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511597 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/88663a9d-6b82-4b98-a1a2-7629c72c28a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511643 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-signing-cabundle\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511690 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3b4d1d22-d5be-42d8-ad04-a8eda187570a-profile-collector-cert\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511831 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511886 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-serving-cert\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511936 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.511986 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5t2g\" (UniqueName: \"kubernetes.io/projected/3b4d1d22-d5be-42d8-ad04-a8eda187570a-kube-api-access-q5t2g\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512053 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c876eadd-e408-42f5-9d74-bf570af2bfae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512099 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-serving-cert\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512163 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3f57836-06b2-4c85-b212-4b260c286b6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512208 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3f57836-06b2-4c85-b212-4b260c286b6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512249 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/72c8ef58-7bca-43c3-900d-297de2e48955-machine-approver-tls\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512316 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-etcd-client\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512349 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512382 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxvsl\" (UniqueName: \"kubernetes.io/projected/72c8ef58-7bca-43c3-900d-297de2e48955-kube-api-access-mxvsl\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512410 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d64d2\" (UniqueName: \"kubernetes.io/projected/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-kube-api-access-d64d2\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512439 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbldt\" (UniqueName: \"kubernetes.io/projected/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-kube-api-access-dbldt\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512469 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148b577d-5144-4ff2-bacf-a8a0fad52c61-config\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512501 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512545 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg7x4\" (UniqueName: \"kubernetes.io/projected/f88009b8-950b-4127-9382-e26e98b95f65-kube-api-access-kg7x4\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512575 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512607 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512671 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-certificates\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512702 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f88009b8-950b-4127-9382-e26e98b95f65-audit-dir\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512737 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-config\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512803 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512867 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-trusted-ca-bundle\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512908 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g65nh\" (UniqueName: \"kubernetes.io/projected/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-kube-api-access-g65nh\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512951 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-service-ca-bundle\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.512995 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-policies\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513042 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cc464dc-c1b0-4306-99ea-bfb435b23a16-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-789nx\" (UID: \"4cc464dc-c1b0-4306-99ea-bfb435b23a16\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513089 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2stp\" (UniqueName: \"kubernetes.io/projected/4cc464dc-c1b0-4306-99ea-bfb435b23a16-kube-api-access-r2stp\") pod \"cluster-samples-operator-665b6dd947-789nx\" (UID: \"4cc464dc-c1b0-4306-99ea-bfb435b23a16\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513312 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-bound-sa-token\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513375 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618b9c5c-956a-4da2-9290-361a423c4e23-serving-cert\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513414 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-dir\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513459 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c876eadd-e408-42f5-9d74-bf570af2bfae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513496 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513528 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e25fa232-dfb3-41d4-a36a-0d5723a67ca3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-46nzg\" (UID: \"e25fa232-dfb3-41d4-a36a-0d5723a67ca3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513637 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876eadd-e408-42f5-9d74-bf570af2bfae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513706 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-audit-policies\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513798 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.513948 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqf64\" (UniqueName: \"kubernetes.io/projected/e3f57836-06b2-4c85-b212-4b260c286b6e-kube-api-access-hqf64\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514023 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-oauth-config\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514133 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514185 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bczf5\" (UniqueName: \"kubernetes.io/projected/618b9c5c-956a-4da2-9290-361a423c4e23-kube-api-access-bczf5\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514278 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db597145-79db-4f18-ae0c-c64f35af2fcc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514340 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514423 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qj77\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-kube-api-access-4qj77\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514473 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c876eadd-e408-42f5-9d74-bf570af2bfae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514521 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-config\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514568 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-service-ca\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514698 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72c8ef58-7bca-43c3-900d-297de2e48955-config\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.514790 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.515519 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3f57836-06b2-4c85-b212-4b260c286b6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.530950 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntnwv\" (UniqueName: \"kubernetes.io/projected/d6d57748-0dc4-4ea8-9da5-2e6488933815-kube-api-access-ntnwv\") pod \"console-operator-58897d9998-mmjq2\" (UID: \"d6d57748-0dc4-4ea8-9da5-2e6488933815\") " pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.548732 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwkg\" (UniqueName: \"kubernetes.io/projected/c289070e-b380-4303-ae1a-b21408ae9f32-kube-api-access-dwwkg\") pod \"apiserver-76f77b778f-c5sq4\" (UID: \"c289070e-b380-4303-ae1a-b21408ae9f32\") " pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.571881 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7btxx\" (UniqueName: \"kubernetes.io/projected/f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee-kube-api-access-7btxx\") pod \"machine-config-controller-84d6567774-xfnmn\" (UID: \"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.581727 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.588385 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58f270a4-ab6a-4810-b99b-2af72f727a15-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.602292 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.625962 4710 request.go:700] Waited for 1.933665628s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/secrets?fieldSelector=metadata.name%3Dcsi-hostpath-provisioner-sa-dockercfg-qd74k&limit=500&resourceVersion=0 Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626428 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626600 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg7x4\" (UniqueName: \"kubernetes.io/projected/f88009b8-950b-4127-9382-e26e98b95f65-kube-api-access-kg7x4\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626633 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626656 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626681 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11deb483-281e-4069-86be-3cc3ebb2f881-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626702 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb507a9b-1015-49a4-8ec3-6f3274511c64-config\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626725 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626765 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76321ee0-be8b-4796-b49c-308aa18e7569-serving-cert\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626789 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2stp\" (UniqueName: \"kubernetes.io/projected/4cc464dc-c1b0-4306-99ea-bfb435b23a16-kube-api-access-r2stp\") pod \"cluster-samples-operator-665b6dd947-789nx\" (UID: \"4cc464dc-c1b0-4306-99ea-bfb435b23a16\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626813 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7d74\" (UniqueName: \"kubernetes.io/projected/730fc09b-f2c9-4cd4-99e6-c60258950823-kube-api-access-f7d74\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626833 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb-metrics-tls\") pod \"dns-operator-744455d44c-5mtxm\" (UID: \"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626853 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g65nh\" (UniqueName: \"kubernetes.io/projected/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-kube-api-access-g65nh\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626874 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cc464dc-c1b0-4306-99ea-bfb435b23a16-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-789nx\" (UID: \"4cc464dc-c1b0-4306-99ea-bfb435b23a16\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626905 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-dir\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626928 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hww5p\" (UniqueName: \"kubernetes.io/projected/f176176a-261c-4a3a-a88f-a8ae63405edf-kube-api-access-hww5p\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626948 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-config\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.626979 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c876eadd-e408-42f5-9d74-bf570af2bfae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627000 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627021 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/69a7d580-fc27-4989-a013-5a30f603fd82-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ff49s\" (UID: \"69a7d580-fc27-4989-a013-5a30f603fd82\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627045 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-registration-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627065 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsp6c\" (UniqueName: \"kubernetes.io/projected/f6866177-110b-4580-b1f6-2b30e5045e08-kube-api-access-rsp6c\") pod \"package-server-manager-789f6589d5-qpnbj\" (UID: \"f6866177-110b-4580-b1f6-2b30e5045e08\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627090 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-audit-policies\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627114 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19675bba-5586-4673-92a9-c930030d8682-apiservice-cert\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627137 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl4hv\" (UniqueName: \"kubernetes.io/projected/11deb483-281e-4069-86be-3cc3ebb2f881-kube-api-access-tl4hv\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627158 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e5240a-2755-47b7-8c93-6c3372379c97-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627178 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-client-ca\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627196 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gnhk\" (UniqueName: \"kubernetes.io/projected/4db3c52f-8fff-4625-ac11-c931baac1233-kube-api-access-5gnhk\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627215 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f62d823-6072-471a-9419-22da5a5af717-cert\") pod \"ingress-canary-fq5v2\" (UID: \"9f62d823-6072-471a-9419-22da5a5af717\") " pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627248 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627268 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25e61509-b211-4f68-b183-52c04179c667-proxy-tls\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627291 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627314 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftj95\" (UniqueName: \"kubernetes.io/projected/c99af3ba-e77e-4b25-ae66-336f74554827-kube-api-access-ftj95\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627334 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627355 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-config-volume\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627377 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db597145-79db-4f18-ae0c-c64f35af2fcc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627401 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627424 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qj77\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-kube-api-access-4qj77\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627443 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c876eadd-e408-42f5-9d74-bf570af2bfae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627462 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-service-ca\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627491 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627513 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25e61509-b211-4f68-b183-52c04179c667-images\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627532 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-config-volume\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627556 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72c8ef58-7bca-43c3-900d-297de2e48955-config\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627587 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3f57836-06b2-4c85-b212-4b260c286b6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627609 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snf64\" (UniqueName: \"kubernetes.io/projected/2d3b8191-b8d2-4d5d-beb8-2390c802d46f-kube-api-access-snf64\") pod \"migrator-59844c95c7-vzxrz\" (UID: \"2d3b8191-b8d2-4d5d-beb8-2390c802d46f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627628 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-trusted-ca\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627649 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c876eadd-e408-42f5-9d74-bf570af2bfae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627669 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c99af3ba-e77e-4b25-ae66-336f74554827-certs\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627701 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/148b577d-5144-4ff2-bacf-a8a0fad52c61-images\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627721 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e5240a-2755-47b7-8c93-6c3372379c97-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627776 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3b4d1d22-d5be-42d8-ad04-a8eda187570a-srv-cert\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627795 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/19675bba-5586-4673-92a9-c930030d8682-tmpfs\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627837 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-mountpoint-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627859 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsp64\" (UniqueName: \"kubernetes.io/projected/148b577d-5144-4ff2-bacf-a8a0fad52c61-kube-api-access-lsp64\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627879 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-encryption-config\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627899 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72c8ef58-7bca-43c3-900d-297de2e48955-auth-proxy-config\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627919 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/730fc09b-f2c9-4cd4-99e6-c60258950823-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627941 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627963 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85137611-745d-417d-98e3-bd3e9285f9c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.627984 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db597145-79db-4f18-ae0c-c64f35af2fcc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628005 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628037 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-config\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628057 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5px9\" (UniqueName: \"kubernetes.io/projected/c68e372e-5db7-475e-bacb-5452d32cbdfc-kube-api-access-p5px9\") pod \"downloads-7954f5f757-m5hbd\" (UID: \"c68e372e-5db7-475e-bacb-5452d32cbdfc\") " pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628082 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/88663a9d-6b82-4b98-a1a2-7629c72c28a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628103 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-signing-cabundle\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628124 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-service-ca\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628143 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djbzw\" (UniqueName: \"kubernetes.io/projected/76321ee0-be8b-4796-b49c-308aa18e7569-kube-api-access-djbzw\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628177 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-ca\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628198 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3b4d1d22-d5be-42d8-ad04-a8eda187570a-profile-collector-cert\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628218 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9e5240a-2755-47b7-8c93-6c3372379c97-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628240 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85137611-745d-417d-98e3-bd3e9285f9c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628272 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628295 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmlxz\" (UniqueName: \"kubernetes.io/projected/c125db0e-2362-4203-88c9-26a1f24c926c-kube-api-access-vmlxz\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628328 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-config\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628349 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628368 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-socket-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628402 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c876eadd-e408-42f5-9d74-bf570af2bfae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628421 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-serving-cert\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628460 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3f57836-06b2-4c85-b212-4b260c286b6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628482 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c99af3ba-e77e-4b25-ae66-336f74554827-node-bootstrap-token\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628515 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w72z\" (UniqueName: \"kubernetes.io/projected/19675bba-5586-4673-92a9-c930030d8682-kube-api-access-7w72z\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628535 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6866177-110b-4580-b1f6-2b30e5045e08-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qpnbj\" (UID: \"f6866177-110b-4580-b1f6-2b30e5045e08\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628569 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/72c8ef58-7bca-43c3-900d-297de2e48955-machine-approver-tls\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628591 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxvsl\" (UniqueName: \"kubernetes.io/projected/72c8ef58-7bca-43c3-900d-297de2e48955-kube-api-access-mxvsl\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628613 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d64d2\" (UniqueName: \"kubernetes.io/projected/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-kube-api-access-d64d2\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628634 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xvpt\" (UniqueName: \"kubernetes.io/projected/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-kube-api-access-2xvpt\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628656 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628680 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cttx5\" (UniqueName: \"kubernetes.io/projected/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-kube-api-access-cttx5\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628738 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-279r4\" (UniqueName: \"kubernetes.io/projected/1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb-kube-api-access-279r4\") pod \"dns-operator-744455d44c-5mtxm\" (UID: \"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628795 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8746\" (UniqueName: \"kubernetes.io/projected/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-kube-api-access-b8746\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628844 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-certificates\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628865 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f88009b8-950b-4127-9382-e26e98b95f65-audit-dir\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628885 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-config\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628909 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz49q\" (UniqueName: \"kubernetes.io/projected/9f62d823-6072-471a-9419-22da5a5af717-kube-api-access-gz49q\") pod \"ingress-canary-fq5v2\" (UID: \"9f62d823-6072-471a-9419-22da5a5af717\") " pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628942 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-trusted-ca-bundle\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628964 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-service-ca-bundle\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.628985 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-policies\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629029 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-bound-sa-token\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629050 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618b9c5c-956a-4da2-9290-361a423c4e23-serving-cert\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629070 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-secret-volume\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629111 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzjz6\" (UniqueName: \"kubernetes.io/projected/2f984d9e-67a3-4e42-b11e-114cb9157256-kube-api-access-wzjz6\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629134 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e25fa232-dfb3-41d4-a36a-0d5723a67ca3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-46nzg\" (UID: \"e25fa232-dfb3-41d4-a36a-0d5723a67ca3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629154 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19675bba-5586-4673-92a9-c930030d8682-webhook-cert\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629175 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876eadd-e408-42f5-9d74-bf570af2bfae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629199 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629225 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-oauth-config\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629246 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f984d9e-67a3-4e42-b11e-114cb9157256-serving-cert\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629278 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-csi-data-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629300 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqf64\" (UniqueName: \"kubernetes.io/projected/e3f57836-06b2-4c85-b212-4b260c286b6e-kube-api-access-hqf64\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629321 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-service-ca-bundle\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629344 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bczf5\" (UniqueName: \"kubernetes.io/projected/618b9c5c-956a-4da2-9290-361a423c4e23-kube-api-access-bczf5\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629374 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-config\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629397 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv4mv\" (UniqueName: \"kubernetes.io/projected/25e61509-b211-4f68-b183-52c04179c667-kube-api-access-xv4mv\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629418 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11deb483-281e-4069-86be-3cc3ebb2f881-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629460 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25e61509-b211-4f68-b183-52c04179c667-auth-proxy-config\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629492 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-client\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629513 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb507a9b-1015-49a4-8ec3-6f3274511c64-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629536 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-tls\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629561 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-signing-key\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629581 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/730fc09b-f2c9-4cd4-99e6-c60258950823-srv-cert\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629612 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f984d9e-67a3-4e42-b11e-114cb9157256-config\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629664 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88663a9d-6b82-4b98-a1a2-7629c72c28a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629691 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42nk\" (UniqueName: \"kubernetes.io/projected/e25fa232-dfb3-41d4-a36a-0d5723a67ca3-kube-api-access-x42nk\") pod \"multus-admission-controller-857f4d67dd-46nzg\" (UID: \"e25fa232-dfb3-41d4-a36a-0d5723a67ca3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629713 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-oauth-serving-cert\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.629736 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/148b577d-5144-4ff2-bacf-a8a0fad52c61-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.631678 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-signing-cabundle\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.631793 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-config\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.631977 4710 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636689 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-metrics-certs\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636786 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb507a9b-1015-49a4-8ec3-6f3274511c64-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636829 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brd76\" (UniqueName: \"kubernetes.io/projected/88663a9d-6b82-4b98-a1a2-7629c72c28a9-kube-api-access-brd76\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636865 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636896 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-default-certificate\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636925 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7w5x\" (UniqueName: \"kubernetes.io/projected/69a7d580-fc27-4989-a013-5a30f603fd82-kube-api-access-q7w5x\") pod \"control-plane-machine-set-operator-78cbb6b69f-ff49s\" (UID: \"69a7d580-fc27-4989-a013-5a30f603fd82\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636953 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-metrics-tls\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.636978 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-plugins-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637045 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-serving-cert\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637070 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637126 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5t2g\" (UniqueName: \"kubernetes.io/projected/3b4d1d22-d5be-42d8-ad04-a8eda187570a-kube-api-access-q5t2g\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637155 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f176176a-261c-4a3a-a88f-a8ae63405edf-serving-cert\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637205 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3f57836-06b2-4c85-b212-4b260c286b6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637233 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-etcd-client\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637261 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637312 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbldt\" (UniqueName: \"kubernetes.io/projected/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-kube-api-access-dbldt\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637340 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5rlg\" (UniqueName: \"kubernetes.io/projected/85137611-745d-417d-98e3-bd3e9285f9c7-kube-api-access-j5rlg\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637367 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148b577d-5144-4ff2-bacf-a8a0fad52c61-config\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637391 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-stats-auth\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.637561 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-audit-policies\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.638032 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.638263 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8q9p\" (UniqueName: \"kubernetes.io/projected/6cdd48c4-7366-4e02-9cf9-78cb66e42249-kube-api-access-q8q9p\") pod \"controller-manager-879f6c89f-gqwlv\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.638687 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-dir\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.639598 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.640392 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.640535 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.640709 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f88009b8-950b-4127-9382-e26e98b95f65-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.641112 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.641393 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c876eadd-e408-42f5-9d74-bf570af2bfae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.641458 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.641851 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c876eadd-e408-42f5-9d74-bf570af2bfae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.641921 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.642187 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72c8ef58-7bca-43c3-900d-297de2e48955-config\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.642203 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-etcd-client\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.642715 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.642868 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-service-ca\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.643029 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cc464dc-c1b0-4306-99ea-bfb435b23a16-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-789nx\" (UID: \"4cc464dc-c1b0-4306-99ea-bfb435b23a16\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.643032 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.643097 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c876eadd-e408-42f5-9d74-bf570af2bfae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.643124 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f88009b8-950b-4127-9382-e26e98b95f65-audit-dir\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.643473 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db597145-79db-4f18-ae0c-c64f35af2fcc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.643493 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-serving-cert\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.644242 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148b577d-5144-4ff2-bacf-a8a0fad52c61-config\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.644377 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.644661 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.645116 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-config\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.645207 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-serving-cert\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.645992 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-certificates\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646052 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-trusted-ca-bundle\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646073 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/148b577d-5144-4ff2-bacf-a8a0fad52c61-images\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646189 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3f57836-06b2-4c85-b212-4b260c286b6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646449 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/88663a9d-6b82-4b98-a1a2-7629c72c28a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646489 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3f57836-06b2-4c85-b212-4b260c286b6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.647171 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-trusted-ca\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.647421 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3b4d1d22-d5be-42d8-ad04-a8eda187570a-profile-collector-cert\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646624 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/618b9c5c-956a-4da2-9290-361a423c4e23-service-ca-bundle\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.646632 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db597145-79db-4f18-ae0c-c64f35af2fcc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.647462 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.647895 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-policies\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.648532 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-oauth-serving-cert\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.649178 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: E1002 12:50:46.649712 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.149657493 +0000 UTC m=+91.256068386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.649730 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f88009b8-950b-4127-9382-e26e98b95f65-encryption-config\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.651434 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3b4d1d22-d5be-42d8-ad04-a8eda187570a-srv-cert\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.651813 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88663a9d-6b82-4b98-a1a2-7629c72c28a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.652162 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/148b577d-5144-4ff2-bacf-a8a0fad52c61-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.652333 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-oauth-config\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.653705 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.653936 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-signing-key\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.654474 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e25fa232-dfb3-41d4-a36a-0d5723a67ca3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-46nzg\" (UID: \"e25fa232-dfb3-41d4-a36a-0d5723a67ca3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.655702 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876eadd-e408-42f5-9d74-bf570af2bfae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.658860 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-tls\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.661616 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618b9c5c-956a-4da2-9290-361a423c4e23-serving-cert\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.667797 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72c8ef58-7bca-43c3-900d-297de2e48955-auth-proxy-config\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.668135 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/72c8ef58-7bca-43c3-900d-297de2e48955-machine-approver-tls\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.674571 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2qz9\" (UniqueName: \"kubernetes.io/projected/58f270a4-ab6a-4810-b99b-2af72f727a15-kube-api-access-g2qz9\") pod \"ingress-operator-5b745b69d9-hj2bl\" (UID: \"58f270a4-ab6a-4810-b99b-2af72f727a15\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.680404 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.694908 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2npwk\" (UniqueName: \"kubernetes.io/projected/90a304b8-7bb2-4d6b-885a-cea681b8d365-kube-api-access-2npwk\") pod \"openshift-apiserver-operator-796bbdcf4f-h97hf\" (UID: \"90a304b8-7bb2-4d6b-885a-cea681b8d365\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.701312 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.721695 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.722857 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.739993 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25e61509-b211-4f68-b183-52c04179c667-images\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740295 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-config-volume\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740322 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snf64\" (UniqueName: \"kubernetes.io/projected/2d3b8191-b8d2-4d5d-beb8-2390c802d46f-kube-api-access-snf64\") pod \"migrator-59844c95c7-vzxrz\" (UID: \"2d3b8191-b8d2-4d5d-beb8-2390c802d46f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740348 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c99af3ba-e77e-4b25-ae66-336f74554827-certs\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740370 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e5240a-2755-47b7-8c93-6c3372379c97-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740400 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/19675bba-5586-4673-92a9-c930030d8682-tmpfs\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740428 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740465 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-mountpoint-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740488 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/730fc09b-f2c9-4cd4-99e6-c60258950823-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740511 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85137611-745d-417d-98e3-bd3e9285f9c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740533 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-config\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740554 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740574 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djbzw\" (UniqueName: \"kubernetes.io/projected/76321ee0-be8b-4796-b49c-308aa18e7569-kube-api-access-djbzw\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740603 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-service-ca\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740627 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9e5240a-2755-47b7-8c93-6c3372379c97-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740647 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-ca\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740676 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85137611-745d-417d-98e3-bd3e9285f9c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740711 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmlxz\" (UniqueName: \"kubernetes.io/projected/c125db0e-2362-4203-88c9-26a1f24c926c-kube-api-access-vmlxz\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740771 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-config\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740801 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740822 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-socket-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740850 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c99af3ba-e77e-4b25-ae66-336f74554827-node-bootstrap-token\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740874 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w72z\" (UniqueName: \"kubernetes.io/projected/19675bba-5586-4673-92a9-c930030d8682-kube-api-access-7w72z\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740896 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6866177-110b-4580-b1f6-2b30e5045e08-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qpnbj\" (UID: \"f6866177-110b-4580-b1f6-2b30e5045e08\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740937 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xvpt\" (UniqueName: \"kubernetes.io/projected/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-kube-api-access-2xvpt\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740963 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cttx5\" (UniqueName: \"kubernetes.io/projected/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-kube-api-access-cttx5\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.740986 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-279r4\" (UniqueName: \"kubernetes.io/projected/1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb-kube-api-access-279r4\") pod \"dns-operator-744455d44c-5mtxm\" (UID: \"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741011 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8746\" (UniqueName: \"kubernetes.io/projected/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-kube-api-access-b8746\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741046 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz49q\" (UniqueName: \"kubernetes.io/projected/9f62d823-6072-471a-9419-22da5a5af717-kube-api-access-gz49q\") pod \"ingress-canary-fq5v2\" (UID: \"9f62d823-6072-471a-9419-22da5a5af717\") " pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741147 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-secret-volume\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741173 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19675bba-5586-4673-92a9-c930030d8682-webhook-cert\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741196 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzjz6\" (UniqueName: \"kubernetes.io/projected/2f984d9e-67a3-4e42-b11e-114cb9157256-kube-api-access-wzjz6\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741228 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f984d9e-67a3-4e42-b11e-114cb9157256-serving-cert\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741261 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-csi-data-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741289 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-service-ca-bundle\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741325 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv4mv\" (UniqueName: \"kubernetes.io/projected/25e61509-b211-4f68-b183-52c04179c667-kube-api-access-xv4mv\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741348 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25e61509-b211-4f68-b183-52c04179c667-auth-proxy-config\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741375 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11deb483-281e-4069-86be-3cc3ebb2f881-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741406 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-client\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741447 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb507a9b-1015-49a4-8ec3-6f3274511c64-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741492 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/730fc09b-f2c9-4cd4-99e6-c60258950823-srv-cert\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741519 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f984d9e-67a3-4e42-b11e-114cb9157256-config\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741564 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb507a9b-1015-49a4-8ec3-6f3274511c64-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741595 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-metrics-certs\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741636 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-default-certificate\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741667 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7w5x\" (UniqueName: \"kubernetes.io/projected/69a7d580-fc27-4989-a013-5a30f603fd82-kube-api-access-q7w5x\") pod \"control-plane-machine-set-operator-78cbb6b69f-ff49s\" (UID: \"69a7d580-fc27-4989-a013-5a30f603fd82\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741694 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-metrics-tls\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741722 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-plugins-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741792 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f176176a-261c-4a3a-a88f-a8ae63405edf-serving-cert\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741846 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5rlg\" (UniqueName: \"kubernetes.io/projected/85137611-745d-417d-98e3-bd3e9285f9c7-kube-api-access-j5rlg\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741876 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-stats-auth\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741905 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11deb483-281e-4069-86be-3cc3ebb2f881-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741948 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb507a9b-1015-49a4-8ec3-6f3274511c64-config\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741977 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76321ee0-be8b-4796-b49c-308aa18e7569-serving-cert\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742027 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7d74\" (UniqueName: \"kubernetes.io/projected/730fc09b-f2c9-4cd4-99e6-c60258950823-kube-api-access-f7d74\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742055 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb-metrics-tls\") pod \"dns-operator-744455d44c-5mtxm\" (UID: \"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742086 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hww5p\" (UniqueName: \"kubernetes.io/projected/f176176a-261c-4a3a-a88f-a8ae63405edf-kube-api-access-hww5p\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742116 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-config\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742144 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsp6c\" (UniqueName: \"kubernetes.io/projected/f6866177-110b-4580-b1f6-2b30e5045e08-kube-api-access-rsp6c\") pod \"package-server-manager-789f6589d5-qpnbj\" (UID: \"f6866177-110b-4580-b1f6-2b30e5045e08\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742188 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/69a7d580-fc27-4989-a013-5a30f603fd82-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ff49s\" (UID: \"69a7d580-fc27-4989-a013-5a30f603fd82\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742266 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-registration-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742303 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl4hv\" (UniqueName: \"kubernetes.io/projected/11deb483-281e-4069-86be-3cc3ebb2f881-kube-api-access-tl4hv\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742342 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19675bba-5586-4673-92a9-c930030d8682-apiservice-cert\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742374 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f62d823-6072-471a-9419-22da5a5af717-cert\") pod \"ingress-canary-fq5v2\" (UID: \"9f62d823-6072-471a-9419-22da5a5af717\") " pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742406 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e5240a-2755-47b7-8c93-6c3372379c97-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742439 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-client-ca\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742470 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gnhk\" (UniqueName: \"kubernetes.io/projected/4db3c52f-8fff-4625-ac11-c931baac1233-kube-api-access-5gnhk\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742500 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742532 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftj95\" (UniqueName: \"kubernetes.io/projected/c99af3ba-e77e-4b25-ae66-336f74554827-kube-api-access-ftj95\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742568 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25e61509-b211-4f68-b183-52c04179c667-proxy-tls\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742599 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742627 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-config-volume\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.743054 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-config\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.744983 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f984d9e-67a3-4e42-b11e-114cb9157256-config\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.745856 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.746705 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-service-ca\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.747365 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-mountpoint-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.747626 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85137611-745d-417d-98e3-bd3e9285f9c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.747682 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-ca\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.748469 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.749287 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-config\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.749362 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e5240a-2755-47b7-8c93-6c3372379c97-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.749839 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/19675bba-5586-4673-92a9-c930030d8682-tmpfs\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: E1002 12:50:46.750121 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.250107495 +0000 UTC m=+91.356518378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.750162 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-registration-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.750297 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-client-ca\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.751154 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c99af3ba-e77e-4b25-ae66-336f74554827-certs\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.751496 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76321ee0-be8b-4796-b49c-308aa18e7569-etcd-client\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.752424 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/730fc09b-f2c9-4cd4-99e6-c60258950823-srv-cert\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.753597 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f62d823-6072-471a-9419-22da5a5af717-cert\") pod \"ingress-canary-fq5v2\" (UID: \"9f62d823-6072-471a-9419-22da5a5af717\") " pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.754207 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76321ee0-be8b-4796-b49c-308aa18e7569-config\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.742145 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-config-volume\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.755332 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-csi-data-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.755846 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-config-volume\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.741693 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25e61509-b211-4f68-b183-52c04179c667-images\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.756322 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-socket-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.756330 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb507a9b-1015-49a4-8ec3-6f3274511c64-config\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.756603 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25e61509-b211-4f68-b183-52c04179c667-proxy-tls\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.756734 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-service-ca-bundle\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.757130 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25e61509-b211-4f68-b183-52c04179c667-auth-proxy-config\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.757294 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11deb483-281e-4069-86be-3cc3ebb2f881-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.757371 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19675bba-5586-4673-92a9-c930030d8682-webhook-cert\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.757483 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4db3c52f-8fff-4625-ac11-c931baac1233-plugins-dir\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.759075 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.759725 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.760151 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/69a7d580-fc27-4989-a013-5a30f603fd82-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ff49s\" (UID: \"69a7d580-fc27-4989-a013-5a30f603fd82\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.762608 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb507a9b-1015-49a4-8ec3-6f3274511c64-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.763399 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/730fc09b-f2c9-4cd4-99e6-c60258950823-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.763952 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e5240a-2755-47b7-8c93-6c3372379c97-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.764416 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.764814 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-metrics-certs\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770185 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85137611-745d-417d-98e3-bd3e9285f9c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770348 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6866177-110b-4580-b1f6-2b30e5045e08-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qpnbj\" (UID: \"f6866177-110b-4580-b1f6-2b30e5045e08\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770508 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c99af3ba-e77e-4b25-ae66-336f74554827-node-bootstrap-token\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770666 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19675bba-5586-4673-92a9-c930030d8682-apiservice-cert\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770671 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-secret-volume\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770713 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f984d9e-67a3-4e42-b11e-114cb9157256-serving-cert\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.770981 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11deb483-281e-4069-86be-3cc3ebb2f881-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.771001 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb-metrics-tls\") pod \"dns-operator-744455d44c-5mtxm\" (UID: \"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.771150 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-metrics-tls\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.771587 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76321ee0-be8b-4796-b49c-308aa18e7569-serving-cert\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.771596 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-default-certificate\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.771836 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.773389 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f176176a-261c-4a3a-a88f-a8ae63405edf-serving-cert\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.791166 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.801837 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.803076 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-stats-auth\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.818684 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.820809 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.843875 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:46 crc kubenswrapper[4710]: E1002 12:50:46.844387 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.344371412 +0000 UTC m=+91.450782295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.859594 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg7x4\" (UniqueName: \"kubernetes.io/projected/f88009b8-950b-4127-9382-e26e98b95f65-kube-api-access-kg7x4\") pod \"apiserver-7bbb656c7d-lj5wt\" (UID: \"f88009b8-950b-4127-9382-e26e98b95f65\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.872715 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn"] Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.874056 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mmjq2"] Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.876005 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brd76\" (UniqueName: \"kubernetes.io/projected/88663a9d-6b82-4b98-a1a2-7629c72c28a9-kube-api-access-brd76\") pod \"openshift-config-operator-7777fb866f-vxtpd\" (UID: \"88663a9d-6b82-4b98-a1a2-7629c72c28a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.896706 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c876eadd-e408-42f5-9d74-bf570af2bfae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9j5lw\" (UID: \"c876eadd-e408-42f5-9d74-bf570af2bfae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.919104 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5t2g\" (UniqueName: \"kubernetes.io/projected/3b4d1d22-d5be-42d8-ad04-a8eda187570a-kube-api-access-q5t2g\") pod \"catalog-operator-68c6474976-wb2db\" (UID: \"3b4d1d22-d5be-42d8-ad04-a8eda187570a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.935332 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2stp\" (UniqueName: \"kubernetes.io/projected/4cc464dc-c1b0-4306-99ea-bfb435b23a16-kube-api-access-r2stp\") pod \"cluster-samples-operator-665b6dd947-789nx\" (UID: \"4cc464dc-c1b0-4306-99ea-bfb435b23a16\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.943900 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gqwlv"] Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.944971 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:46 crc kubenswrapper[4710]: E1002 12:50:46.945394 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.445374678 +0000 UTC m=+91.551785621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.956966 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g65nh\" (UniqueName: \"kubernetes.io/projected/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-kube-api-access-g65nh\") pod \"console-f9d7485db-mrtmb\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:46 crc kubenswrapper[4710]: W1002 12:50:46.963617 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cdd48c4_7366_4e02_9cf9_78cb66e42249.slice/crio-a13c1c9604f1f15c6aaf4ff7845cc434030bef0f510b96ae69dcd02c97164616 WatchSource:0}: Error finding container a13c1c9604f1f15c6aaf4ff7845cc434030bef0f510b96ae69dcd02c97164616: Status 404 returned error can't find the container with id a13c1c9604f1f15c6aaf4ff7845cc434030bef0f510b96ae69dcd02c97164616 Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.976240 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3f57836-06b2-4c85-b212-4b260c286b6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.981584 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf"] Oct 02 12:50:46 crc kubenswrapper[4710]: W1002 12:50:46.993107 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90a304b8_7bb2_4d6b_885a_cea681b8d365.slice/crio-ab3c89b27808020379a044d113ce55a5adce68146225083eacc4126572e2667c WatchSource:0}: Error finding container ab3c89b27808020379a044d113ce55a5adce68146225083eacc4126572e2667c: Status 404 returned error can't find the container with id ab3c89b27808020379a044d113ce55a5adce68146225083eacc4126572e2667c Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.996147 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" Oct 02 12:50:46 crc kubenswrapper[4710]: I1002 12:50:46.996263 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d64d2\" (UniqueName: \"kubernetes.io/projected/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-kube-api-access-d64d2\") pod \"oauth-openshift-558db77b4-284s4\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.002560 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.008277 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.017451 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbldt\" (UniqueName: \"kubernetes.io/projected/b95e2431-460e-4bac-a4ce-61fb7b6b44aa-kube-api-access-dbldt\") pod \"service-ca-9c57cc56f-b7kkk\" (UID: \"b95e2431-460e-4bac-a4ce-61fb7b6b44aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.018478 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c5sq4"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.034415 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qj77\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-kube-api-access-4qj77\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.045737 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.045876 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.545854911 +0000 UTC m=+91.652265794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.046248 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.046588 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.546579889 +0000 UTC m=+91.652990772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.053010 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxvsl\" (UniqueName: \"kubernetes.io/projected/72c8ef58-7bca-43c3-900d-297de2e48955-kube-api-access-mxvsl\") pod \"machine-approver-56656f9798-gp652\" (UID: \"72c8ef58-7bca-43c3-900d-297de2e48955\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.065171 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.074084 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsp64\" (UniqueName: \"kubernetes.io/projected/148b577d-5144-4ff2-bacf-a8a0fad52c61-kube-api-access-lsp64\") pod \"machine-api-operator-5694c8668f-hq8zf\" (UID: \"148b577d-5144-4ff2-bacf-a8a0fad52c61\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.080525 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.095626 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bczf5\" (UniqueName: \"kubernetes.io/projected/618b9c5c-956a-4da2-9290-361a423c4e23-kube-api-access-bczf5\") pod \"authentication-operator-69f744f599-9cmjq\" (UID: \"618b9c5c-956a-4da2-9290-361a423c4e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.115072 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqf64\" (UniqueName: \"kubernetes.io/projected/e3f57836-06b2-4c85-b212-4b260c286b6e-kube-api-access-hqf64\") pod \"cluster-image-registry-operator-dc59b4c8b-5xwm4\" (UID: \"e3f57836-06b2-4c85-b212-4b260c286b6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.123484 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.125400 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58f270a4_ab6a_4810_b99b_2af72f727a15.slice/crio-ee3e68d9a51836b0b8f232553579c8d8d977f9f8ca56e0e88eea54a77a202f9b WatchSource:0}: Error finding container ee3e68d9a51836b0b8f232553579c8d8d977f9f8ca56e0e88eea54a77a202f9b: Status 404 returned error can't find the container with id ee3e68d9a51836b0b8f232553579c8d8d977f9f8ca56e0e88eea54a77a202f9b Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.131012 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.141613 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc289070e_b380_4303_ae1a_b21408ae9f32.slice/crio-e07440a8e3729eb1dabca44a4594bfc09df0996fae08f77f3e11f94d7f063634 WatchSource:0}: Error finding container e07440a8e3729eb1dabca44a4594bfc09df0996fae08f77f3e11f94d7f063634: Status 404 returned error can't find the container with id e07440a8e3729eb1dabca44a4594bfc09df0996fae08f77f3e11f94d7f063634 Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.143395 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-bound-sa-token\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.147635 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.147778 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.647726938 +0000 UTC m=+91.754137831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.147945 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.148384 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.648369095 +0000 UTC m=+91.754779998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.155005 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.169516 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5px9\" (UniqueName: \"kubernetes.io/projected/c68e372e-5db7-475e-bacb-5452d32cbdfc-kube-api-access-p5px9\") pod \"downloads-7954f5f757-m5hbd\" (UID: \"c68e372e-5db7-475e-bacb-5452d32cbdfc\") " pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.182071 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.182851 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42nk\" (UniqueName: \"kubernetes.io/projected/e25fa232-dfb3-41d4-a36a-0d5723a67ca3-kube-api-access-x42nk\") pod \"multus-admission-controller-857f4d67dd-46nzg\" (UID: \"e25fa232-dfb3-41d4-a36a-0d5723a67ca3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.195672 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snf64\" (UniqueName: \"kubernetes.io/projected/2d3b8191-b8d2-4d5d-beb8-2390c802d46f-kube-api-access-snf64\") pod \"migrator-59844c95c7-vzxrz\" (UID: \"2d3b8191-b8d2-4d5d-beb8-2390c802d46f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.210040 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.220204 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmlxz\" (UniqueName: \"kubernetes.io/projected/c125db0e-2362-4203-88c9-26a1f24c926c-kube-api-access-vmlxz\") pod \"marketplace-operator-79b997595-cfvqq\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.236539 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb507a9b-1015-49a4-8ec3-6f3274511c64-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-skcgv\" (UID: \"eb507a9b-1015-49a4-8ec3-6f3274511c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.248716 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.248974 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.74894848 +0000 UTC m=+91.855359363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.249300 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.249644 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.749637447 +0000 UTC m=+91.856048330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.257148 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hww5p\" (UniqueName: \"kubernetes.io/projected/f176176a-261c-4a3a-a88f-a8ae63405edf-kube-api-access-hww5p\") pod \"route-controller-manager-6576b87f9c-ftnw6\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.259017 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.274346 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f00b704d-ef9c-4bbd-a29e-16ff6e6d374d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jq4b6\" (UID: \"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.280603 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.290438 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.296522 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djbzw\" (UniqueName: \"kubernetes.io/projected/76321ee0-be8b-4796-b49c-308aa18e7569-kube-api-access-djbzw\") pod \"etcd-operator-b45778765-c7t49\" (UID: \"76321ee0-be8b-4796-b49c-308aa18e7569\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.317870 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9e5240a-2755-47b7-8c93-6c3372379c97-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgc6q\" (UID: \"b9e5240a-2755-47b7-8c93-6c3372379c97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.318132 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.322948 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-284s4"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.325969 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.333708 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.337366 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8746\" (UniqueName: \"kubernetes.io/projected/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-kube-api-access-b8746\") pod \"collect-profiles-29323485-9pxkz\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.339927 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.345260 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.345698 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.350675 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.351280 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.851259919 +0000 UTC m=+91.957670802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.354172 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gnhk\" (UniqueName: \"kubernetes.io/projected/4db3c52f-8fff-4625-ac11-c931baac1233-kube-api-access-5gnhk\") pod \"csi-hostpathplugin-xmcpk\" (UID: \"4db3c52f-8fff-4625-ac11-c931baac1233\") " pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.373244 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.375047 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl4hv\" (UniqueName: \"kubernetes.io/projected/11deb483-281e-4069-86be-3cc3ebb2f881-kube-api-access-tl4hv\") pod \"kube-storage-version-migrator-operator-b67b599dd-mwmsd\" (UID: \"11deb483-281e-4069-86be-3cc3ebb2f881\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.380203 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.395893 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz49q\" (UniqueName: \"kubernetes.io/projected/9f62d823-6072-471a-9419-22da5a5af717-kube-api-access-gz49q\") pod \"ingress-canary-fq5v2\" (UID: \"9f62d823-6072-471a-9419-22da5a5af717\") " pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.406577 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc876eadd_e408_42f5_9d74_bf570af2bfae.slice/crio-d95377b2bf8255725a17768ca5ccb041872ce0353289565e893bdee8a0a56c9b WatchSource:0}: Error finding container d95377b2bf8255725a17768ca5ccb041872ce0353289565e893bdee8a0a56c9b: Status 404 returned error can't find the container with id d95377b2bf8255725a17768ca5ccb041872ce0353289565e893bdee8a0a56c9b Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.417705 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xvpt\" (UniqueName: \"kubernetes.io/projected/7ed87807-1891-43c5-9a11-d70c8fdb6ffe-kube-api-access-2xvpt\") pod \"dns-default-xt66p\" (UID: \"7ed87807-1891-43c5-9a11-d70c8fdb6ffe\") " pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.433862 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.436315 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cttx5\" (UniqueName: \"kubernetes.io/projected/c8b54dd0-afe8-4f80-ab3d-db6040735d5a-kube-api-access-cttx5\") pod \"router-default-5444994796-j786v\" (UID: \"c8b54dd0-afe8-4f80-ab3d-db6040735d5a\") " pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.438800 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.452835 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.453190 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:47.953172607 +0000 UTC m=+92.059583490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.453436 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-279r4\" (UniqueName: \"kubernetes.io/projected/1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb-kube-api-access-279r4\") pod \"dns-operator-744455d44c-5mtxm\" (UID: \"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.469712 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.474625 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.475270 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzjz6\" (UniqueName: \"kubernetes.io/projected/2f984d9e-67a3-4e42-b11e-114cb9157256-kube-api-access-wzjz6\") pod \"service-ca-operator-777779d784-49mwb\" (UID: \"2f984d9e-67a3-4e42-b11e-114cb9157256\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.481866 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf88009b8_950b_4127_9382_e26e98b95f65.slice/crio-601d3278cf4d0ee9ba4679747479be7c42c96a6b43c2b69bf476c76af03d9ae5 WatchSource:0}: Error finding container 601d3278cf4d0ee9ba4679747479be7c42c96a6b43c2b69bf476c76af03d9ae5: Status 404 returned error can't find the container with id 601d3278cf4d0ee9ba4679747479be7c42c96a6b43c2b69bf476c76af03d9ae5 Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.495733 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.496100 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5rlg\" (UniqueName: \"kubernetes.io/projected/85137611-745d-417d-98e3-bd3e9285f9c7-kube-api-access-j5rlg\") pod \"openshift-controller-manager-operator-756b6f6bc6-pqll9\" (UID: \"85137611-745d-417d-98e3-bd3e9285f9c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.496105 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" event={"ID":"90a304b8-7bb2-4d6b-885a-cea681b8d365","Type":"ContainerStarted","Data":"ab3c89b27808020379a044d113ce55a5adce68146225083eacc4126572e2667c"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.497017 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" event={"ID":"d6d57748-0dc4-4ea8-9da5-2e6488933815","Type":"ContainerStarted","Data":"0d0b67614498a097e6f5cebd83ca9adf3e75fef2595f40a8378511826a31626c"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.497802 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" event={"ID":"c876eadd-e408-42f5-9d74-bf570af2bfae","Type":"ContainerStarted","Data":"d95377b2bf8255725a17768ca5ccb041872ce0353289565e893bdee8a0a56c9b"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.499411 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" event={"ID":"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d","Type":"ContainerStarted","Data":"36eee9a1a5e9a7990a5ed8c085df27fc30ea349f0f50f229278427e08b23299e"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.500524 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" event={"ID":"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee","Type":"ContainerStarted","Data":"0dcc832be5f989d286932282df0119775072a9981ff369b29b9ba106f30bf723"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.517003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" event={"ID":"88663a9d-6b82-4b98-a1a2-7629c72c28a9","Type":"ContainerStarted","Data":"f4cdadfc0216c5ab5722d7a93e9a85fd40fed06c0358f52cf249b1ab0173337b"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.519346 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" event={"ID":"c289070e-b380-4303-ae1a-b21408ae9f32","Type":"ContainerStarted","Data":"e07440a8e3729eb1dabca44a4594bfc09df0996fae08f77f3e11f94d7f063634"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.521284 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsp6c\" (UniqueName: \"kubernetes.io/projected/f6866177-110b-4580-b1f6-2b30e5045e08-kube-api-access-rsp6c\") pod \"package-server-manager-789f6589d5-qpnbj\" (UID: \"f6866177-110b-4580-b1f6-2b30e5045e08\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.521417 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" event={"ID":"f88009b8-950b-4127-9382-e26e98b95f65","Type":"ContainerStarted","Data":"601d3278cf4d0ee9ba4679747479be7c42c96a6b43c2b69bf476c76af03d9ae5"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.524517 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" event={"ID":"6cdd48c4-7366-4e02-9cf9-78cb66e42249","Type":"ContainerStarted","Data":"a13c1c9604f1f15c6aaf4ff7845cc434030bef0f510b96ae69dcd02c97164616"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.526582 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" event={"ID":"58f270a4-ab6a-4810-b99b-2af72f727a15","Type":"ContainerStarted","Data":"ee3e68d9a51836b0b8f232553579c8d8d977f9f8ca56e0e88eea54a77a202f9b"} Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.529040 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.537068 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7d74\" (UniqueName: \"kubernetes.io/projected/730fc09b-f2c9-4cd4-99e6-c60258950823-kube-api-access-f7d74\") pod \"olm-operator-6b444d44fb-bllq5\" (UID: \"730fc09b-f2c9-4cd4-99e6-c60258950823\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.544335 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.554191 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.554563 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.054513342 +0000 UTC m=+92.160924235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.554864 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.555324 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.055308822 +0000 UTC m=+92.161719705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.561233 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w72z\" (UniqueName: \"kubernetes.io/projected/19675bba-5586-4673-92a9-c930030d8682-kube-api-access-7w72z\") pod \"packageserver-d55dfcdfc-qrllc\" (UID: \"19675bba-5586-4673-92a9-c930030d8682\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.581284 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.583460 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv4mv\" (UniqueName: \"kubernetes.io/projected/25e61509-b211-4f68-b183-52c04179c667-kube-api-access-xv4mv\") pod \"machine-config-operator-74547568cd-26qxb\" (UID: \"25e61509-b211-4f68-b183-52c04179c667\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.592072 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.603638 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftj95\" (UniqueName: \"kubernetes.io/projected/c99af3ba-e77e-4b25-ae66-336f74554827-kube-api-access-ftj95\") pod \"machine-config-server-vpc2k\" (UID: \"c99af3ba-e77e-4b25-ae66-336f74554827\") " pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.603723 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.604807 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.615475 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.615677 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7w5x\" (UniqueName: \"kubernetes.io/projected/69a7d580-fc27-4989-a013-5a30f603fd82-kube-api-access-q7w5x\") pod \"control-plane-machine-set-operator-78cbb6b69f-ff49s\" (UID: \"69a7d580-fc27-4989-a013-5a30f603fd82\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.648368 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fq5v2" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.655162 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vpc2k" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.655610 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.656137 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.156116123 +0000 UTC m=+92.262527006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.661873 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.729995 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-c7t49"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.747503 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.749602 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.771108 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.771139 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.771656 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.271642953 +0000 UTC m=+92.378053836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.790053 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.797428 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.856566 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hq8zf"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.863993 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mrtmb"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.870493 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.871219 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb507a9b_1015_49a4_8ec3_6f3274511c64.slice/crio-1ba732a7086035ae41b73cfcb830dd75a04ba568b0260b1ebe56a37651ea5c94 WatchSource:0}: Error finding container 1ba732a7086035ae41b73cfcb830dd75a04ba568b0260b1ebe56a37651ea5c94: Status 404 returned error can't find the container with id 1ba732a7086035ae41b73cfcb830dd75a04ba568b0260b1ebe56a37651ea5c94 Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.872194 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.872605 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.372590338 +0000 UTC m=+92.479001221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.876108 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4"] Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.912183 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6"] Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.949492 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ef751b2_7d8e_4caa_ad61_d8544a501c6b.slice/crio-2035507cfec1479c3405989148a50bafda2d592b692195b1f1a8163e53d7db71 WatchSource:0}: Error finding container 2035507cfec1479c3405989148a50bafda2d592b692195b1f1a8163e53d7db71: Status 404 returned error can't find the container with id 2035507cfec1479c3405989148a50bafda2d592b692195b1f1a8163e53d7db71 Oct 02 12:50:47 crc kubenswrapper[4710]: I1002 12:50:47.976654 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:47 crc kubenswrapper[4710]: E1002 12:50:47.977050 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.477027579 +0000 UTC m=+92.583438462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:47 crc kubenswrapper[4710]: W1002 12:50:47.991128 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3f57836_06b2_4c85_b212_4b260c286b6e.slice/crio-9a9f9eda335cc975ed55dc2f8b23e9c562005b70753e11dcc4f0c40c297b327b WatchSource:0}: Error finding container 9a9f9eda335cc975ed55dc2f8b23e9c562005b70753e11dcc4f0c40c297b327b: Status 404 returned error can't find the container with id 9a9f9eda335cc975ed55dc2f8b23e9c562005b70753e11dcc4f0c40c297b327b Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.056808 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xmcpk"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.078634 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.078910 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.578881346 +0000 UTC m=+92.685292229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.080516 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.080864 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.580854086 +0000 UTC m=+92.687264969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.181267 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.181602 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.681567564 +0000 UTC m=+92.787978447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.226479 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-m5hbd"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.242338 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-9cmjq"] Oct 02 12:50:48 crc kubenswrapper[4710]: W1002 12:50:48.248681 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4db3c52f_8fff_4625_ac11_c931baac1233.slice/crio-fc92f685778646328283c421a3fa20488d4c84b1dda7b094a23afa7cb12fc0ae WatchSource:0}: Error finding container fc92f685778646328283c421a3fa20488d4c84b1dda7b094a23afa7cb12fc0ae: Status 404 returned error can't find the container with id fc92f685778646328283c421a3fa20488d4c84b1dda7b094a23afa7cb12fc0ae Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.282462 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.282857 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.782839547 +0000 UTC m=+92.889250430 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.386378 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.386797 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.886782266 +0000 UTC m=+92.993193149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.455203 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.490736 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.491036 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:48.991024554 +0000 UTC m=+93.097435437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.505438 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.526629 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.533150 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b7kkk"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.546430 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cfvqq"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.552350 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.554781 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.579874 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" event={"ID":"3b4d1d22-d5be-42d8-ad04-a8eda187570a","Type":"ContainerStarted","Data":"0e1fc116887c0c38da798e140af38f1623bccfbb4b19f67398866325a5346ad6"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.585119 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" event={"ID":"72c8ef58-7bca-43c3-900d-297de2e48955","Type":"ContainerStarted","Data":"f6e306c7b7822a62815388fa559c9f6ab8f9b943e4b5e9da4304ea427023840f"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.588084 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" event={"ID":"148b577d-5144-4ff2-bacf-a8a0fad52c61","Type":"ContainerStarted","Data":"83a63bcb5f7ed679a5bf5f6225060263f3a6e317424f276f61eda5e1d8cb5fad"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.591254 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.591584 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.091570618 +0000 UTC m=+93.197981491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.591870 4710 generic.go:334] "Generic (PLEG): container finished" podID="c289070e-b380-4303-ae1a-b21408ae9f32" containerID="74b86c96e4ba5add2378f60f2626f21d0501c1154704c3fc6fd33764c3ee235e" exitCode=0 Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.591922 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" event={"ID":"c289070e-b380-4303-ae1a-b21408ae9f32","Type":"ContainerDied","Data":"74b86c96e4ba5add2378f60f2626f21d0501c1154704c3fc6fd33764c3ee235e"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.595796 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" event={"ID":"4cc464dc-c1b0-4306-99ea-bfb435b23a16","Type":"ContainerStarted","Data":"48bdf531e7f53cac616b5bacc98b3a0621e651501ec6436ede752922ee880d92"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.601467 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" event={"ID":"6cdd48c4-7366-4e02-9cf9-78cb66e42249","Type":"ContainerStarted","Data":"e32ab99cfa7e73c0a4994afc8b467e0231928e090472e8c9b6c8e4d1062824e0"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.601831 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.604405 4710 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-gqwlv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.604438 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" podUID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.607719 4710 generic.go:334] "Generic (PLEG): container finished" podID="88663a9d-6b82-4b98-a1a2-7629c72c28a9" containerID="a9fc445d3b78c36a1e79b0ecf14e702f4278682f971a3a2f0face9affa8f6e4f" exitCode=0 Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.607802 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" event={"ID":"88663a9d-6b82-4b98-a1a2-7629c72c28a9","Type":"ContainerDied","Data":"a9fc445d3b78c36a1e79b0ecf14e702f4278682f971a3a2f0face9affa8f6e4f"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.623670 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" event={"ID":"d6d57748-0dc4-4ea8-9da5-2e6488933815","Type":"ContainerStarted","Data":"2ab79d7acfee5bf18453346ec3abdeb15df55b80b7a4d9ec5d0ccdaf1ae62a1e"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.623985 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.628491 4710 patch_prober.go:28] interesting pod/console-operator-58897d9998-mmjq2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.628543 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" podUID="d6d57748-0dc4-4ea8-9da5-2e6488933815" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.630328 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.637017 4710 generic.go:334] "Generic (PLEG): container finished" podID="f88009b8-950b-4127-9382-e26e98b95f65" containerID="584bfa3a8acf82559714990e6545f13fe6a4a2622cb1e1f23f63bcd471d812d0" exitCode=0 Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.637103 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" event={"ID":"f88009b8-950b-4127-9382-e26e98b95f65","Type":"ContainerDied","Data":"584bfa3a8acf82559714990e6545f13fe6a4a2622cb1e1f23f63bcd471d812d0"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.639188 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" event={"ID":"76321ee0-be8b-4796-b49c-308aa18e7569","Type":"ContainerStarted","Data":"5242d3c6cdcde3ef20c61041968d60bb70da72e4933b7d6af74c429a774df56c"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.641281 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" event={"ID":"618b9c5c-956a-4da2-9290-361a423c4e23","Type":"ContainerStarted","Data":"839b17d1fc4b2cb7fd73dcaeb6ef82175025a6c5188d2a97049ffd225a4269d7"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.643688 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" event={"ID":"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d","Type":"ContainerStarted","Data":"ba99bdaa823d37dc1177a2d212530b87cd79407ebdd695aea591733b71c2b4dc"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.644406 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.644477 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.646681 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" event={"ID":"58f270a4-ab6a-4810-b99b-2af72f727a15","Type":"ContainerStarted","Data":"1871b5e3f4bbfab8564684df6ef86855a367ea178dd6dd03caa912282c44cb09"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.648546 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" event={"ID":"90a304b8-7bb2-4d6b-885a-cea681b8d365","Type":"ContainerStarted","Data":"4961299fa6c961f3c86e8a1d9e4395d39b39f004ccb55285682b73fb945fd52b"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.652341 4710 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-284s4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.652391 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" podUID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.656479 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" event={"ID":"c876eadd-e408-42f5-9d74-bf570af2bfae","Type":"ContainerStarted","Data":"08ea8a62f7d2ca47b901b3cc8ea8519fb45c10d3e4bd09c061d10f1cd623a435"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.659583 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" event={"ID":"eb507a9b-1015-49a4-8ec3-6f3274511c64","Type":"ContainerStarted","Data":"1ba732a7086035ae41b73cfcb830dd75a04ba568b0260b1ebe56a37651ea5c94"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.661136 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" event={"ID":"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee","Type":"ContainerStarted","Data":"648b9d9e892f68e67047d1cb408be5d08c004d5e5d5b625d74cb0963dfe2bce1"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.662222 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mrtmb" event={"ID":"1ef751b2-7d8e-4caa-ad61-d8544a501c6b","Type":"ContainerStarted","Data":"2035507cfec1479c3405989148a50bafda2d592b692195b1f1a8163e53d7db71"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.663484 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" event={"ID":"f176176a-261c-4a3a-a88f-a8ae63405edf","Type":"ContainerStarted","Data":"ab2eaa0649551d0b24f625ec782731e826cb4f51a207d0552ba7a0e571505fb2"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.666186 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vpc2k" event={"ID":"c99af3ba-e77e-4b25-ae66-336f74554827","Type":"ContainerStarted","Data":"a888538aff485a15f857b1246751a79d214a362c8c3414272ffe1e9969b85a16"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.668035 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-j786v" event={"ID":"c8b54dd0-afe8-4f80-ab3d-db6040735d5a","Type":"ContainerStarted","Data":"789144b87b669a3c2a2277c66a50610321276534ea9d50b87634ba4119686562"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.670793 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" event={"ID":"4db3c52f-8fff-4625-ac11-c931baac1233","Type":"ContainerStarted","Data":"fc92f685778646328283c421a3fa20488d4c84b1dda7b094a23afa7cb12fc0ae"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.682994 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-46nzg"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.685668 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-m5hbd" event={"ID":"c68e372e-5db7-475e-bacb-5452d32cbdfc","Type":"ContainerStarted","Data":"6a0ac4b83ff5857db0c7b1af9690de8595f1a8c1bcbfc6f3f481ba8eab8fce29"} Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.692631 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.695197 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.195181889 +0000 UTC m=+93.301592782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.703848 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" event={"ID":"e3f57836-06b2-4c85-b212-4b260c286b6e","Type":"ContainerStarted","Data":"9a9f9eda335cc975ed55dc2f8b23e9c562005b70753e11dcc4f0c40c297b327b"} Oct 02 12:50:48 crc kubenswrapper[4710]: W1002 12:50:48.747686 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d3b8191_b8d2_4d5d_beb8_2390c802d46f.slice/crio-ada54ccf5c18d442ce1b38ecc23e0cef65abf8f923907a0e76908a7c383282d1 WatchSource:0}: Error finding container ada54ccf5c18d442ce1b38ecc23e0cef65abf8f923907a0e76908a7c383282d1: Status 404 returned error can't find the container with id ada54ccf5c18d442ce1b38ecc23e0cef65abf8f923907a0e76908a7c383282d1 Oct 02 12:50:48 crc kubenswrapper[4710]: W1002 12:50:48.759440 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc125db0e_2362_4203_88c9_26a1f24c926c.slice/crio-686bab809f0704f3ecea6a0b7d17bf203a8bac1795fc840c90e45322cd06435a WatchSource:0}: Error finding container 686bab809f0704f3ecea6a0b7d17bf203a8bac1795fc840c90e45322cd06435a: Status 404 returned error can't find the container with id 686bab809f0704f3ecea6a0b7d17bf203a8bac1795fc840c90e45322cd06435a Oct 02 12:50:48 crc kubenswrapper[4710]: W1002 12:50:48.766718 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85137611_745d_417d_98e3_bd3e9285f9c7.slice/crio-8da92f9647c6b7576ee9106a4827e19b344f80fa8e638a42f1df42f8f0396f43 WatchSource:0}: Error finding container 8da92f9647c6b7576ee9106a4827e19b344f80fa8e638a42f1df42f8f0396f43: Status 404 returned error can't find the container with id 8da92f9647c6b7576ee9106a4827e19b344f80fa8e638a42f1df42f8f0396f43 Oct 02 12:50:48 crc kubenswrapper[4710]: W1002 12:50:48.783161 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf00b704d_ef9c_4bbd_a29e_16ff6e6d374d.slice/crio-9120b35948b56d29dc5d8587bc653604b6e0b61bc37ef1ef1e6f4ce7326189e0 WatchSource:0}: Error finding container 9120b35948b56d29dc5d8587bc653604b6e0b61bc37ef1ef1e6f4ce7326189e0: Status 404 returned error can't find the container with id 9120b35948b56d29dc5d8587bc653604b6e0b61bc37ef1ef1e6f4ce7326189e0 Oct 02 12:50:48 crc kubenswrapper[4710]: W1002 12:50:48.790729 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11deb483_281e_4069_86be_3cc3ebb2f881.slice/crio-56f430c1ba3c68e2f2243bb8ca2199e34ab82c7981869b9b66f30e7ba9029e95 WatchSource:0}: Error finding container 56f430c1ba3c68e2f2243bb8ca2199e34ab82c7981869b9b66f30e7ba9029e95: Status 404 returned error can't find the container with id 56f430c1ba3c68e2f2243bb8ca2199e34ab82c7981869b9b66f30e7ba9029e95 Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.793272 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.793772 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.293725753 +0000 UTC m=+93.400136646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.794309 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.801941 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.301661623 +0000 UTC m=+93.408072506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.895117 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.895781 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.395736815 +0000 UTC m=+93.502147698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.895970 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.896262 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.396249738 +0000 UTC m=+93.502660621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.945435 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" podStartSLOduration=71.945418272 podStartE2EDuration="1m11.945418272s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:48.941526594 +0000 UTC m=+93.047937477" watchObservedRunningTime="2025-10-02 12:50:48.945418272 +0000 UTC m=+93.051829155" Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.993600 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj"] Oct 02 12:50:48 crc kubenswrapper[4710]: I1002 12:50:48.998003 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:48 crc kubenswrapper[4710]: E1002 12:50:48.998310 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.49829186 +0000 UTC m=+93.604702743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.037293 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6866177_110b_4580_b1f6_2b30e5045e08.slice/crio-f3bb634318a59dac1a008bb91aa7d2a7b29ad685aed60cc44b41160422656616 WatchSource:0}: Error finding container f3bb634318a59dac1a008bb91aa7d2a7b29ad685aed60cc44b41160422656616: Status 404 returned error can't find the container with id f3bb634318a59dac1a008bb91aa7d2a7b29ad685aed60cc44b41160422656616 Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.040277 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz"] Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.043153 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb"] Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.047062 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xt66p"] Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.100227 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.100851 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.600839534 +0000 UTC m=+93.707250417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.101018 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25e61509_b211_4f68_b183_52c04179c667.slice/crio-09d5d2c8f2381127d4f813b82179689e0af42c3b7da42da7aa0c960d9c4a8aa4 WatchSource:0}: Error finding container 09d5d2c8f2381127d4f813b82179689e0af42c3b7da42da7aa0c960d9c4a8aa4: Status 404 returned error can't find the container with id 09d5d2c8f2381127d4f813b82179689e0af42c3b7da42da7aa0c960d9c4a8aa4 Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.109433 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s"] Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.112522 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5mtxm"] Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.130701 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fq5v2"] Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.131665 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e3d7fcf_5054_4b16_a92f_ca3b9b3bd9eb.slice/crio-c9261e0a4acaf75920142327ed727765afe69b88ac8f722650ffc562b78eb55c WatchSource:0}: Error finding container c9261e0a4acaf75920142327ed727765afe69b88ac8f722650ffc562b78eb55c: Status 404 returned error can't find the container with id c9261e0a4acaf75920142327ed727765afe69b88ac8f722650ffc562b78eb55c Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.135295 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ed87807_1891_43c5_9a11_d70c8fdb6ffe.slice/crio-348d329bf7049f5a491df503fd8c45cdf80e4a3b133bf672831d0b996965e305 WatchSource:0}: Error finding container 348d329bf7049f5a491df503fd8c45cdf80e4a3b133bf672831d0b996965e305: Status 404 returned error can't find the container with id 348d329bf7049f5a491df503fd8c45cdf80e4a3b133bf672831d0b996965e305 Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.143064 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-49mwb"] Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.213306 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.213658 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.713625096 +0000 UTC m=+93.820035979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.213851 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.214257 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.714247432 +0000 UTC m=+93.820658315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.226869 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9j5lw" podStartSLOduration=72.226846018 podStartE2EDuration="1m12.226846018s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:49.177732785 +0000 UTC m=+93.284143658" watchObservedRunningTime="2025-10-02 12:50:49.226846018 +0000 UTC m=+93.333256901" Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.238152 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69a7d580_fc27_4989_a013_5a30f603fd82.slice/crio-dfab839bb6e911638f6c2d05cff149d312051a2b458a97824159f8873fb3ee3a WatchSource:0}: Error finding container dfab839bb6e911638f6c2d05cff149d312051a2b458a97824159f8873fb3ee3a: Status 404 returned error can't find the container with id dfab839bb6e911638f6c2d05cff149d312051a2b458a97824159f8873fb3ee3a Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.253950 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f62d823_6072_471a_9419_22da5a5af717.slice/crio-5f2c4ba6270a091ab1fc58df971fd661472898276ca8e1e34afc74713739ec45 WatchSource:0}: Error finding container 5f2c4ba6270a091ab1fc58df971fd661472898276ca8e1e34afc74713739ec45: Status 404 returned error can't find the container with id 5f2c4ba6270a091ab1fc58df971fd661472898276ca8e1e34afc74713739ec45 Oct 02 12:50:49 crc kubenswrapper[4710]: W1002 12:50:49.259481 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f984d9e_67a3_4e42_b11e_114cb9157256.slice/crio-0d7c7e06bb92dafd06e9d21e19ee5f032dd4cc3d48d147f7430513c2e01ec291 WatchSource:0}: Error finding container 0d7c7e06bb92dafd06e9d21e19ee5f032dd4cc3d48d147f7430513c2e01ec291: Status 404 returned error can't find the container with id 0d7c7e06bb92dafd06e9d21e19ee5f032dd4cc3d48d147f7430513c2e01ec291 Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.322211 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.322345 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.822315955 +0000 UTC m=+93.928726838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.323327 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.323934 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.823915025 +0000 UTC m=+93.930325908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.425035 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.425449 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:49.925432894 +0000 UTC m=+94.031843777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.499036 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" podStartSLOduration=72.499020411 podStartE2EDuration="1m12.499020411s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:49.49858217 +0000 UTC m=+93.604993053" watchObservedRunningTime="2025-10-02 12:50:49.499020411 +0000 UTC m=+93.605431294" Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.532188 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.532599 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.032585884 +0000 UTC m=+94.138996767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.633468 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.633581 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.133560889 +0000 UTC m=+94.239971772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.640817 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.641308 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.141297634 +0000 UTC m=+94.247708517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.741403 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.741624 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.241610212 +0000 UTC m=+94.348021095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.758385 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" event={"ID":"3b4d1d22-d5be-42d8-ad04-a8eda187570a","Type":"ContainerStarted","Data":"810ba4b18566e225c8efcdf70fd0d5eff2d41ac8b30e36c7dcc91c14c6526b0a"} Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.759382 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.763382 4710 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-wb2db container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.763426 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" podUID="3b4d1d22-d5be-42d8-ad04-a8eda187570a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.782954 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h97hf" podStartSLOduration=72.782927629 podStartE2EDuration="1m12.782927629s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:49.78057449 +0000 UTC m=+93.886985373" watchObservedRunningTime="2025-10-02 12:50:49.782927629 +0000 UTC m=+93.889338512" Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.828894 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" event={"ID":"148b577d-5144-4ff2-bacf-a8a0fad52c61","Type":"ContainerStarted","Data":"a93eaae3bdfb982f8b6c3931e3bb4d6b5d9afe393afc9253c54d84eea38d17cf"} Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.828946 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" event={"ID":"148b577d-5144-4ff2-bacf-a8a0fad52c61","Type":"ContainerStarted","Data":"378d5564de40b9741cfdb3f631e614ceb4ed49fbc57e747f0ed2a63b5c283e17"} Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.862439 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.863736 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.363721258 +0000 UTC m=+94.470132141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.963504 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:49 crc kubenswrapper[4710]: E1002 12:50:49.964053 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.464038607 +0000 UTC m=+94.570449490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:49 crc kubenswrapper[4710]: I1002 12:50:49.997458 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" podStartSLOduration=72.997436335 podStartE2EDuration="1m12.997436335s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:49.941427849 +0000 UTC m=+94.047838732" watchObservedRunningTime="2025-10-02 12:50:49.997436335 +0000 UTC m=+94.103847218" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.000111 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" event={"ID":"e25fa232-dfb3-41d4-a36a-0d5723a67ca3","Type":"ContainerStarted","Data":"035aa92bc031e138eb648af115db354f40420febef698ebbd32aad493d99b30a"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.037352 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mrtmb" event={"ID":"1ef751b2-7d8e-4caa-ad61-d8544a501c6b","Type":"ContainerStarted","Data":"dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.052012 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" podStartSLOduration=73.051989495 podStartE2EDuration="1m13.051989495s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.041218954 +0000 UTC m=+94.147629827" watchObservedRunningTime="2025-10-02 12:50:50.051989495 +0000 UTC m=+94.158400378" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.068817 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.069375 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.569358281 +0000 UTC m=+94.675769164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.078286 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" event={"ID":"c125db0e-2362-4203-88c9-26a1f24c926c","Type":"ContainerStarted","Data":"c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.078378 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" event={"ID":"c125db0e-2362-4203-88c9-26a1f24c926c","Type":"ContainerStarted","Data":"686bab809f0704f3ecea6a0b7d17bf203a8bac1795fc840c90e45322cd06435a"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.079691 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.092993 4710 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cfvqq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.093081 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.115200 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" event={"ID":"69a7d580-fc27-4989-a013-5a30f603fd82","Type":"ContainerStarted","Data":"dfab839bb6e911638f6c2d05cff149d312051a2b458a97824159f8873fb3ee3a"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.139782 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" event={"ID":"25e61509-b211-4f68-b183-52c04179c667","Type":"ContainerStarted","Data":"4fbfcbc2410ef29c2d674f45707fec26427b23ee3e8045871bfc593104b08c74"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.139836 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" event={"ID":"25e61509-b211-4f68-b183-52c04179c667","Type":"ContainerStarted","Data":"09d5d2c8f2381127d4f813b82179689e0af42c3b7da42da7aa0c960d9c4a8aa4"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.143577 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mrtmb" podStartSLOduration=73.143561834 podStartE2EDuration="1m13.143561834s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.103369595 +0000 UTC m=+94.209780478" watchObservedRunningTime="2025-10-02 12:50:50.143561834 +0000 UTC m=+94.249972717" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.147460 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" podStartSLOduration=73.147448652 podStartE2EDuration="1m13.147448652s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.144256591 +0000 UTC m=+94.250667474" watchObservedRunningTime="2025-10-02 12:50:50.147448652 +0000 UTC m=+94.253859535" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.150220 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" event={"ID":"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d","Type":"ContainerStarted","Data":"9120b35948b56d29dc5d8587bc653604b6e0b61bc37ef1ef1e6f4ce7326189e0"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.159029 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vpc2k" event={"ID":"c99af3ba-e77e-4b25-ae66-336f74554827","Type":"ContainerStarted","Data":"56a886f6e4b0dd51e8e11890692cd591d801a4b0f4c02a70f3168d548947ef7f"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.180903 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" podStartSLOduration=73.180879401 podStartE2EDuration="1m13.180879401s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.178191673 +0000 UTC m=+94.284602556" watchObservedRunningTime="2025-10-02 12:50:50.180879401 +0000 UTC m=+94.287290304" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.182730 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.182854 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.68283548 +0000 UTC m=+94.789246363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.183735 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.184099 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.684082361 +0000 UTC m=+94.790493244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.195319 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" event={"ID":"11deb483-281e-4069-86be-3cc3ebb2f881","Type":"ContainerStarted","Data":"2eae24a5b0b4c8c91d84c19909446e8b3476787d1292b54d1a3c1e40add40023"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.195377 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" event={"ID":"11deb483-281e-4069-86be-3cc3ebb2f881","Type":"ContainerStarted","Data":"56f430c1ba3c68e2f2243bb8ca2199e34ab82c7981869b9b66f30e7ba9029e95"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.213737 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" event={"ID":"e3f57836-06b2-4c85-b212-4b260c286b6e","Type":"ContainerStarted","Data":"95976b262f15d817a7f9d4132bb50ccf56988fa423cf6e5c242430cf43899856"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.224177 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" event={"ID":"72c8ef58-7bca-43c3-900d-297de2e48955","Type":"ContainerStarted","Data":"bffd4aad63179fcb1d75cf2e6ad9975b9c6b091c364da798f85b3a6a736ad09d"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.252546 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" podStartSLOduration=73.25252292 podStartE2EDuration="1m13.25252292s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.206431922 +0000 UTC m=+94.312842805" watchObservedRunningTime="2025-10-02 12:50:50.25252292 +0000 UTC m=+94.358933803" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.255452 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" event={"ID":"4cc464dc-c1b0-4306-99ea-bfb435b23a16","Type":"ContainerStarted","Data":"ae1525d1cccc2b6af98e5d4d4c02df191fb20f023358634f8e06bcce4f21eeb8"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.255487 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" event={"ID":"4cc464dc-c1b0-4306-99ea-bfb435b23a16","Type":"ContainerStarted","Data":"15336579409ddf7874041f725b6899c810e393884b6631b94901102e243761a4"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.260503 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vpc2k" podStartSLOduration=6.260481459 podStartE2EDuration="6.260481459s" podCreationTimestamp="2025-10-02 12:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.229459381 +0000 UTC m=+94.335870264" watchObservedRunningTime="2025-10-02 12:50:50.260481459 +0000 UTC m=+94.366892332" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.283470 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" event={"ID":"b95e2431-460e-4bac-a4ce-61fb7b6b44aa","Type":"ContainerStarted","Data":"d294204516ca43c7dd1c3c46cc2bef8561a53e7b6d89d536b1464f6eaaa51620"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.283519 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" event={"ID":"b95e2431-460e-4bac-a4ce-61fb7b6b44aa","Type":"ContainerStarted","Data":"7467461d2e741c6572d4b8aeb16e7bb792eeab2f18b00fbfa7eb3f0ebceccf9c"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.285125 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.285390 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.785376974 +0000 UTC m=+94.891787857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.289172 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xwm4" podStartSLOduration=73.289146629 podStartE2EDuration="1m13.289146629s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.253641388 +0000 UTC m=+94.360052271" watchObservedRunningTime="2025-10-02 12:50:50.289146629 +0000 UTC m=+94.395557512" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.297233 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mwmsd" podStartSLOduration=73.297212062 podStartE2EDuration="1m13.297212062s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.277157448 +0000 UTC m=+94.383568321" watchObservedRunningTime="2025-10-02 12:50:50.297212062 +0000 UTC m=+94.403622945" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.324139 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" event={"ID":"19675bba-5586-4673-92a9-c930030d8682","Type":"ContainerStarted","Data":"b5f2da1a61144899a417846f2a1496a6abdc0baae69ba2b98048db8a7175273a"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.324183 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" event={"ID":"19675bba-5586-4673-92a9-c930030d8682","Type":"ContainerStarted","Data":"55060589373bf0dc533fae1c82da369cb9b6c3c567fd029812122123ce1d9e0d"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.324769 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.334937 4710 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qrllc container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.335117 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" podUID="19675bba-5586-4673-92a9-c930030d8682" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.355409 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" event={"ID":"c289070e-b380-4303-ae1a-b21408ae9f32","Type":"ContainerStarted","Data":"a71958d471d507b66d7a769749b25a041b8420fa0c602681687c4699467b5e54"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.356097 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-b7kkk" podStartSLOduration=72.35608084 podStartE2EDuration="1m12.35608084s" podCreationTimestamp="2025-10-02 12:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.355631918 +0000 UTC m=+94.462042811" watchObservedRunningTime="2025-10-02 12:50:50.35608084 +0000 UTC m=+94.462491723" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.356531 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-789nx" podStartSLOduration=73.356525741 podStartE2EDuration="1m13.356525741s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.314702971 +0000 UTC m=+94.421113864" watchObservedRunningTime="2025-10-02 12:50:50.356525741 +0000 UTC m=+94.462936624" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.391026 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-m5hbd" event={"ID":"c68e372e-5db7-475e-bacb-5452d32cbdfc","Type":"ContainerStarted","Data":"ded0e24b3cfd20b23355b16d1e83f7e62f1cebb0eb615beed9f6f3c6b8e311e8"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.391880 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.393729 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.395809 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.895796357 +0000 UTC m=+95.002207250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.409656 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.409989 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.411844 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" event={"ID":"730fc09b-f2c9-4cd4-99e6-c60258950823","Type":"ContainerStarted","Data":"9ef6c71b494cdaad8c224b4adfa3a75b144f110414ca2f50a6c077baa6accbe5"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.411877 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.411886 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" event={"ID":"730fc09b-f2c9-4cd4-99e6-c60258950823","Type":"ContainerStarted","Data":"3c236f511be2b2f3a1be4bcfeedddfb1000eb0fea01246137b2b99941460b402"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.419725 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" podStartSLOduration=72.419707987 podStartE2EDuration="1m12.419707987s" podCreationTimestamp="2025-10-02 12:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.392878634 +0000 UTC m=+94.499289517" watchObservedRunningTime="2025-10-02 12:50:50.419707987 +0000 UTC m=+94.526118870" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.428638 4710 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bllq5 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.428680 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" podUID="730fc09b-f2c9-4cd4-99e6-c60258950823" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.450939 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" event={"ID":"b9e5240a-2755-47b7-8c93-6c3372379c97","Type":"ContainerStarted","Data":"8ff2bce9d35cf17e092ba548020a6d72e9110e6bdbadc5e6bea93bf659f184d0"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.450984 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" event={"ID":"b9e5240a-2755-47b7-8c93-6c3372379c97","Type":"ContainerStarted","Data":"2370189db348d8e4dbd49dc640da474989978fe8fda66f41646e249ecd763b2d"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.471162 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-m5hbd" podStartSLOduration=73.471146249 podStartE2EDuration="1m13.471146249s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.421627245 +0000 UTC m=+94.528038118" watchObservedRunningTime="2025-10-02 12:50:50.471146249 +0000 UTC m=+94.577557132" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.472535 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" podStartSLOduration=73.472529573 podStartE2EDuration="1m13.472529573s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.469129368 +0000 UTC m=+94.575540261" watchObservedRunningTime="2025-10-02 12:50:50.472529573 +0000 UTC m=+94.578940456" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.482256 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-j786v" event={"ID":"c8b54dd0-afe8-4f80-ab3d-db6040735d5a","Type":"ContainerStarted","Data":"b03e2dc0b915d1b229e5d2e04a134f78b2876ec5f58f5646647ced6c8b3822c7"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.491942 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" event={"ID":"618b9c5c-956a-4da2-9290-361a423c4e23","Type":"ContainerStarted","Data":"4f3c206a9e2b4c989fb7b5f46a58cafdaa8d27dd77cf56f4c0a389c4e1b1b71f"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.494456 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.494785 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:50.994772332 +0000 UTC m=+95.101183215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.497533 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" event={"ID":"2d3b8191-b8d2-4d5d-beb8-2390c802d46f","Type":"ContainerStarted","Data":"38b9601a9c14a2ab8eaf66cc09dfd06db3eef3c6328ebaf9e2192935a7ad3ae5"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.497570 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" event={"ID":"2d3b8191-b8d2-4d5d-beb8-2390c802d46f","Type":"ContainerStarted","Data":"ada54ccf5c18d442ce1b38ecc23e0cef65abf8f923907a0e76908a7c383282d1"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.505034 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fq5v2" event={"ID":"9f62d823-6072-471a-9419-22da5a5af717","Type":"ContainerStarted","Data":"5f2c4ba6270a091ab1fc58df971fd661472898276ca8e1e34afc74713739ec45"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.520195 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" event={"ID":"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb","Type":"ContainerStarted","Data":"c9261e0a4acaf75920142327ed727765afe69b88ac8f722650ffc562b78eb55c"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.527300 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" event={"ID":"2f984d9e-67a3-4e42-b11e-114cb9157256","Type":"ContainerStarted","Data":"0d7c7e06bb92dafd06e9d21e19ee5f032dd4cc3d48d147f7430513c2e01ec291"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.535410 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" podStartSLOduration=73.535394042 podStartE2EDuration="1m13.535394042s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.534847348 +0000 UTC m=+94.641258241" watchObservedRunningTime="2025-10-02 12:50:50.535394042 +0000 UTC m=+94.641804925" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.536299 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgc6q" podStartSLOduration=73.536294524 podStartE2EDuration="1m13.536294524s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.504986968 +0000 UTC m=+94.611397851" watchObservedRunningTime="2025-10-02 12:50:50.536294524 +0000 UTC m=+94.642705407" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.541530 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" event={"ID":"f88009b8-950b-4127-9382-e26e98b95f65","Type":"ContainerStarted","Data":"bfaf83487ae7b23281e4f89283f5bce02a48a49cde48286a8f16aa0321904e96"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.546585 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" event={"ID":"85137611-745d-417d-98e3-bd3e9285f9c7","Type":"ContainerStarted","Data":"eee9b318e89172c3f1550930baeb6955699d08056864adaf891b170d09692839"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.546620 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" event={"ID":"85137611-745d-417d-98e3-bd3e9285f9c7","Type":"ContainerStarted","Data":"8da92f9647c6b7576ee9106a4827e19b344f80fa8e638a42f1df42f8f0396f43"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.549380 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" event={"ID":"eb507a9b-1015-49a4-8ec3-6f3274511c64","Type":"ContainerStarted","Data":"fb545691efdc8d943fa95a476b861d4a703e45e94ceec968811ad58bcdfb2875"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.555510 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" event={"ID":"f6866177-110b-4580-b1f6-2b30e5045e08","Type":"ContainerStarted","Data":"6e9353bb8182a7855e9c790fdcbec3bdbff934d422551d0762a0e7347d28b160"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.555554 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" event={"ID":"f6866177-110b-4580-b1f6-2b30e5045e08","Type":"ContainerStarted","Data":"f3bb634318a59dac1a008bb91aa7d2a7b29ad685aed60cc44b41160422656616"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.555779 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.564383 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" event={"ID":"f16d0d3d-79ed-4ce0-ba45-f119eeb2ffee","Type":"ContainerStarted","Data":"c525bb5e1f72e8b377e8268e5f56122e6e98d32e844d55096275804a3021b2f5"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.595656 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.596418 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.096400113 +0000 UTC m=+95.202810996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.598924 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" event={"ID":"58f270a4-ab6a-4810-b99b-2af72f727a15","Type":"ContainerStarted","Data":"e62576c630f250754c64bd4b4db1171bfa78e66c9dabaeef0cbf3fffa525a988"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.609923 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-j786v" podStartSLOduration=73.609907902 podStartE2EDuration="1m13.609907902s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.608421735 +0000 UTC m=+94.714832618" watchObservedRunningTime="2025-10-02 12:50:50.609907902 +0000 UTC m=+94.716318785" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.611356 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-9cmjq" podStartSLOduration=73.611246516 podStartE2EDuration="1m13.611246516s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.570467702 +0000 UTC m=+94.676878585" watchObservedRunningTime="2025-10-02 12:50:50.611246516 +0000 UTC m=+94.717657399" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.611781 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" event={"ID":"f176176a-261c-4a3a-a88f-a8ae63405edf","Type":"ContainerStarted","Data":"8e9f15d78a87dcd63369cb886a1ccb184b9b6a43a31bcd13e75a829184bb5a58"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.612466 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.642421 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xt66p" event={"ID":"7ed87807-1891-43c5-9a11-d70c8fdb6ffe","Type":"ContainerStarted","Data":"348d329bf7049f5a491df503fd8c45cdf80e4a3b133bf672831d0b996965e305"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.690315 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" event={"ID":"76321ee0-be8b-4796-b49c-308aa18e7569","Type":"ContainerStarted","Data":"cbbbbdddc19cc932b5b93bb58e680ce2ed7e279fda042448d76bd8a96cdab2cf"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.693305 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" event={"ID":"88663a9d-6b82-4b98-a1a2-7629c72c28a9","Type":"ContainerStarted","Data":"e1f67a3a0f00f361a3cea3311ec1445dcad992c5974a239b0372c5a706e35d49"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.693791 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.696475 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.697788 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.197739208 +0000 UTC m=+95.304178711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.714559 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" event={"ID":"e8205f2e-3067-4ab0-9e30-43d99dfd53eb","Type":"ContainerStarted","Data":"2a31f35a0c7ea30b986761fb35e2f9a1337cbe48a6da9fe6bd96035bc3c7c59c"} Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.726976 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.733849 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.737197 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mmjq2" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.748472 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xfnmn" podStartSLOduration=73.748456471 podStartE2EDuration="1m13.748456471s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.672242538 +0000 UTC m=+94.778653421" watchObservedRunningTime="2025-10-02 12:50:50.748456471 +0000 UTC m=+94.854867354" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.771480 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.793277 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" podStartSLOduration=73.793262006 podStartE2EDuration="1m13.793262006s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.791351418 +0000 UTC m=+94.897762301" watchObservedRunningTime="2025-10-02 12:50:50.793262006 +0000 UTC m=+94.899672889" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.793373 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pqll9" podStartSLOduration=73.793368569 podStartE2EDuration="1m13.793368569s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.752126343 +0000 UTC m=+94.858537226" watchObservedRunningTime="2025-10-02 12:50:50.793368569 +0000 UTC m=+94.899779452" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.799597 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.804173 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.30415851 +0000 UTC m=+95.410569393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.833802 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" podStartSLOduration=72.833783883 podStartE2EDuration="1m12.833783883s" podCreationTimestamp="2025-10-02 12:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.832191193 +0000 UTC m=+94.938602076" watchObservedRunningTime="2025-10-02 12:50:50.833783883 +0000 UTC m=+94.940194766" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.870201 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hj2bl" podStartSLOduration=73.870187367 podStartE2EDuration="1m13.870187367s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.86911735 +0000 UTC m=+94.975528233" watchObservedRunningTime="2025-10-02 12:50:50.870187367 +0000 UTC m=+94.976598250" Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.901841 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:50 crc kubenswrapper[4710]: E1002 12:50:50.926258 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.426236755 +0000 UTC m=+95.532647638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:50 crc kubenswrapper[4710]: I1002 12:50:50.973161 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-skcgv" podStartSLOduration=73.973146282 podStartE2EDuration="1m13.973146282s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.93005077 +0000 UTC m=+95.036461663" watchObservedRunningTime="2025-10-02 12:50:50.973146282 +0000 UTC m=+95.079557165" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.004130 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.004378 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.504364266 +0000 UTC m=+95.610775149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.016107 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-fq5v2" podStartSLOduration=7.01608999 podStartE2EDuration="7.01608999s" podCreationTimestamp="2025-10-02 12:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.015561477 +0000 UTC m=+95.121972360" watchObservedRunningTime="2025-10-02 12:50:51.01608999 +0000 UTC m=+95.122500873" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.016312 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" podStartSLOduration=74.016309126 podStartE2EDuration="1m14.016309126s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:50.975394559 +0000 UTC m=+95.081805442" watchObservedRunningTime="2025-10-02 12:50:51.016309126 +0000 UTC m=+95.122719999" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.080955 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" podStartSLOduration=73.080897658 podStartE2EDuration="1m13.080897658s" podCreationTimestamp="2025-10-02 12:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.066048885 +0000 UTC m=+95.172459768" watchObservedRunningTime="2025-10-02 12:50:51.080897658 +0000 UTC m=+95.187308541" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.105272 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.105607 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.605592278 +0000 UTC m=+95.712003161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.181625 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" podStartSLOduration=74.181604566 podStartE2EDuration="1m14.181604566s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.128195205 +0000 UTC m=+95.234606098" watchObservedRunningTime="2025-10-02 12:50:51.181604566 +0000 UTC m=+95.288015459" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.206359 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.206681 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.706670465 +0000 UTC m=+95.813081348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.307903 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.308114 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.808083442 +0000 UTC m=+95.914494335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.308245 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.308511 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.808499902 +0000 UTC m=+95.914910785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.343587 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-c7t49" podStartSLOduration=74.343572393 podStartE2EDuration="1m14.343572393s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.256152298 +0000 UTC m=+95.362563181" watchObservedRunningTime="2025-10-02 12:50:51.343572393 +0000 UTC m=+95.449983276" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.408703 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.408838 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.908817041 +0000 UTC m=+96.015227924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.409266 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.409560 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:51.909552779 +0000 UTC m=+96.015963652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.475507 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.479478 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:51 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:51 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:51 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.479532 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.510313 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.510521 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.010498963 +0000 UTC m=+96.116909846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.510625 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.510899 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.010891843 +0000 UTC m=+96.117302726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.611271 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.611672 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.111649842 +0000 UTC m=+96.218060725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.712533 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.713076 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.213056308 +0000 UTC m=+96.319467241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.736453 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-49mwb" event={"ID":"2f984d9e-67a3-4e42-b11e-114cb9157256","Type":"ContainerStarted","Data":"a7dd827c2fb8f4940d9ceae0a7dfec17f4256fec046db14345e484d142852e82"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.737541 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" event={"ID":"4db3c52f-8fff-4625-ac11-c931baac1233","Type":"ContainerStarted","Data":"e4bdb9c1071ec00bb61adb925b11d458f85919e4d0f124ac8b56a5fcf25f4091"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.739640 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" event={"ID":"e8205f2e-3067-4ab0-9e30-43d99dfd53eb","Type":"ContainerStarted","Data":"615d81488b245ad7e8fb43787023141b4ba3fc1d197076bceafe1fe7a4195f68"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.742297 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" event={"ID":"72c8ef58-7bca-43c3-900d-297de2e48955","Type":"ContainerStarted","Data":"0c3b22da94f437bf47ec58bf0ba99b1e92bde8f9ca958e334759fddf6005f9c9"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.744816 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" event={"ID":"e25fa232-dfb3-41d4-a36a-0d5723a67ca3","Type":"ContainerStarted","Data":"0ad973fee69380444a7e3a1b6d9cbe6415ab9cbc261da00f048f741bf18ec0f5"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.744861 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" event={"ID":"e25fa232-dfb3-41d4-a36a-0d5723a67ca3","Type":"ContainerStarted","Data":"335bc21d939630b04972c69f4ab67f60d285e91ebe35dfd9fa7ce26b5660d60c"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.746804 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ff49s" event={"ID":"69a7d580-fc27-4989-a013-5a30f603fd82","Type":"ContainerStarted","Data":"68deb6230ea13888da164f96c2cf96df99b5a1ac8490d48bf0d790843f646b19"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.749626 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" event={"ID":"f6866177-110b-4580-b1f6-2b30e5045e08","Type":"ContainerStarted","Data":"daf1039c03b53b1fa13e4dc78446fb7c1d83bf29cda6eca5421acba942da0295"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.752124 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xt66p" event={"ID":"7ed87807-1891-43c5-9a11-d70c8fdb6ffe","Type":"ContainerStarted","Data":"9a8fe6b0e4cf0bee268b7618678700c5adecf78c6a9cbeba1aec1bcc7d32ea53"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.752156 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xt66p" event={"ID":"7ed87807-1891-43c5-9a11-d70c8fdb6ffe","Type":"ContainerStarted","Data":"0b478f2f1509a9c2e0345e48f4bfeaf0f5d1ddc1ce1a60f9a26c41892e258616"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.752591 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-xt66p" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.758995 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fq5v2" event={"ID":"9f62d823-6072-471a-9419-22da5a5af717","Type":"ContainerStarted","Data":"34cbe4d993313a3ecb354ee70d622f1a56f19d9615d84e90f171e102ed3fc909"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.762022 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" event={"ID":"25e61509-b211-4f68-b183-52c04179c667","Type":"ContainerStarted","Data":"874d91751862a6ffd7454a926e494c5c19d0d20e959cd79732b28eb4f868fcd4"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.765645 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jq4b6" event={"ID":"f00b704d-ef9c-4bbd-a29e-16ff6e6d374d","Type":"ContainerStarted","Data":"2c2a0ec6b2593ef87d671495afdbf1486457efea13a1550e1951da38b6cf5607"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.775494 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" event={"ID":"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb","Type":"ContainerStarted","Data":"be81daa17d8e85986b6ca3eb06b30815a230b7d4ce82636328bc770bf06ad429"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.775559 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" event={"ID":"1e3d7fcf-5054-4b16-a92f-ca3b9b3bd9eb","Type":"ContainerStarted","Data":"a7f2065a87c29b1b9ad327bbb77b5b699b1faf53495d2013b65259004cc33b0e"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.779145 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" event={"ID":"c289070e-b380-4303-ae1a-b21408ae9f32","Type":"ContainerStarted","Data":"e1f29dc8f28e59526b695a3f0fb6bdeceb06632d9415b325a1b1945aa78d4630"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.784559 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vzxrz" event={"ID":"2d3b8191-b8d2-4d5d-beb8-2390c802d46f","Type":"ContainerStarted","Data":"6620f5745b64fef9dc32fb8b777d4ccea659a7ffb149e3b7728d50a980266c47"} Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.785000 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.785042 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.785670 4710 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cfvqq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.785699 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.793077 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wb2db" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.800770 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bllq5" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.813941 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.814117 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.314089365 +0000 UTC m=+96.420500248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.814363 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.814718 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.314710791 +0000 UTC m=+96.421121674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.818779 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.819015 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.819944 4710 patch_prober.go:28] interesting pod/apiserver-76f77b778f-c5sq4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.31:8443/livez\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.819983 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" podUID="c289070e-b380-4303-ae1a-b21408ae9f32" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.31:8443/livez\": dial tcp 10.217.0.31:8443: connect: connection refused" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.832504 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" podStartSLOduration=74.832483877 podStartE2EDuration="1m14.832483877s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.487555667 +0000 UTC m=+95.593966550" watchObservedRunningTime="2025-10-02 12:50:51.832483877 +0000 UTC m=+95.938894760" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.832617 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gp652" podStartSLOduration=74.8326121 podStartE2EDuration="1m14.8326121s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.83099968 +0000 UTC m=+95.937410563" watchObservedRunningTime="2025-10-02 12:50:51.8326121 +0000 UTC m=+95.939022983" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.908573 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-hq8zf" podStartSLOduration=74.908550127 podStartE2EDuration="1m14.908550127s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:51.905777647 +0000 UTC m=+96.012188530" watchObservedRunningTime="2025-10-02 12:50:51.908550127 +0000 UTC m=+96.014961010" Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.915534 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.915701 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.415677775 +0000 UTC m=+96.522088658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:51 crc kubenswrapper[4710]: I1002 12:50:51.925080 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:51 crc kubenswrapper[4710]: E1002 12:50:51.930549 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.430529838 +0000 UTC m=+96.536940721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.054336 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.054798 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.554780278 +0000 UTC m=+96.661191161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.124254 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.124315 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.145934 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-xt66p" podStartSLOduration=8.145913946 podStartE2EDuration="8.145913946s" podCreationTimestamp="2025-10-02 12:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:52.142630834 +0000 UTC m=+96.249041717" watchObservedRunningTime="2025-10-02 12:50:52.145913946 +0000 UTC m=+96.252324829" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.156085 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.156532 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.656519692 +0000 UTC m=+96.762930575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.172807 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-46nzg" podStartSLOduration=75.172793401 podStartE2EDuration="1m15.172793401s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:52.170855812 +0000 UTC m=+96.277266695" watchObservedRunningTime="2025-10-02 12:50:52.172793401 +0000 UTC m=+96.279204284" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.257708 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.258186 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.758171275 +0000 UTC m=+96.864582158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.280083 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" podStartSLOduration=75.280043114 podStartE2EDuration="1m15.280043114s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:52.215000751 +0000 UTC m=+96.321411634" watchObservedRunningTime="2025-10-02 12:50:52.280043114 +0000 UTC m=+96.386453987" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.328369 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-5mtxm" podStartSLOduration=75.328351917 podStartE2EDuration="1m15.328351917s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:52.281113941 +0000 UTC m=+96.387524824" watchObservedRunningTime="2025-10-02 12:50:52.328351917 +0000 UTC m=+96.434762800" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.329320 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-26qxb" podStartSLOduration=75.329313211 podStartE2EDuration="1m15.329313211s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:52.325725571 +0000 UTC m=+96.432136454" watchObservedRunningTime="2025-10-02 12:50:52.329313211 +0000 UTC m=+96.435724094" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.359239 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.359725 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.859710244 +0000 UTC m=+96.966121127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.453218 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kkj24"] Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.454534 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.460049 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.460452 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:52.960436373 +0000 UTC m=+97.066847256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.460924 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.478579 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kkj24"] Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.484999 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:52 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:52 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:52 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.485085 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.562036 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-utilities\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.562097 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.562189 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pvsm\" (UniqueName: \"kubernetes.io/projected/8527e0f4-d514-45e1-89db-2d07c359387d-kube-api-access-8pvsm\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.562209 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-catalog-content\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.562550 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.062523876 +0000 UTC m=+97.168934759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.636889 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7tgfs"] Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.637935 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.643142 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.651695 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7tgfs"] Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663201 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663427 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-catalog-content\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663469 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-utilities\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663518 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-utilities\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663609 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhsw8\" (UniqueName: \"kubernetes.io/projected/567598d7-6c35-46b0-bd80-5c04b4f7e024-kube-api-access-bhsw8\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663648 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pvsm\" (UniqueName: \"kubernetes.io/projected/8527e0f4-d514-45e1-89db-2d07c359387d-kube-api-access-8pvsm\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.663676 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-catalog-content\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.664204 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-catalog-content\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.664299 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.164278161 +0000 UTC m=+97.270689044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.664649 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-utilities\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.680710 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pvsm\" (UniqueName: \"kubernetes.io/projected/8527e0f4-d514-45e1-89db-2d07c359387d-kube-api-access-8pvsm\") pod \"certified-operators-kkj24\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.712004 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.765399 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-catalog-content\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.765449 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.765472 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-utilities\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.765532 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhsw8\" (UniqueName: \"kubernetes.io/projected/567598d7-6c35-46b0-bd80-5c04b4f7e024-kube-api-access-bhsw8\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.766247 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-catalog-content\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.766526 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.266514328 +0000 UTC m=+97.372925211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.766971 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-utilities\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.783593 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.784007 4710 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qrllc container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.784074 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" podUID="19675bba-5586-4673-92a9-c930030d8682" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.804131 4710 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cfvqq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.804175 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.804230 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.804248 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.808587 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhsw8\" (UniqueName: \"kubernetes.io/projected/567598d7-6c35-46b0-bd80-5c04b4f7e024-kube-api-access-bhsw8\") pod \"community-operators-7tgfs\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.818883 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lj5wt" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.856111 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tzk59"] Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.894779 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:52 crc kubenswrapper[4710]: E1002 12:50:52.897112 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.397095246 +0000 UTC m=+97.503506129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.913646 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.950295 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.958699 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tzk59"] Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.997781 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwr5z\" (UniqueName: \"kubernetes.io/projected/2ef28114-3b05-44a8-b1e5-5bbbda54d657-kube-api-access-hwr5z\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.997858 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-utilities\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.997877 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-catalog-content\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:52 crc kubenswrapper[4710]: I1002 12:50:52.997899 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:52.998246 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.498231435 +0000 UTC m=+97.604642318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.042124 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kxfsf"] Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.043549 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.066362 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxfsf"] Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.098663 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.098897 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-catalog-content\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.098962 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hpw6\" (UniqueName: \"kubernetes.io/projected/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-kube-api-access-6hpw6\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.099069 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwr5z\" (UniqueName: \"kubernetes.io/projected/2ef28114-3b05-44a8-b1e5-5bbbda54d657-kube-api-access-hwr5z\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.099123 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-catalog-content\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.099148 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-utilities\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.099168 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-utilities\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.099277 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.599259262 +0000 UTC m=+97.705670145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.099707 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-catalog-content\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.100009 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-utilities\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.147049 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwr5z\" (UniqueName: \"kubernetes.io/projected/2ef28114-3b05-44a8-b1e5-5bbbda54d657-kube-api-access-hwr5z\") pod \"certified-operators-tzk59\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.202710 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-catalog-content\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.202788 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-utilities\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.202836 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.202860 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hpw6\" (UniqueName: \"kubernetes.io/projected/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-kube-api-access-6hpw6\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.203928 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-catalog-content\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.204312 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-utilities\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.204678 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.704662058 +0000 UTC m=+97.811072941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.227284 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hpw6\" (UniqueName: \"kubernetes.io/projected/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-kube-api-access-6hpw6\") pod \"community-operators-kxfsf\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.258484 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.307161 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.307570 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.807554692 +0000 UTC m=+97.913965575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.343732 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7tgfs"] Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.352021 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vxtpd" Oct 02 12:50:53 crc kubenswrapper[4710]: W1002 12:50:53.352312 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod567598d7_6c35_46b0_bd80_5c04b4f7e024.slice/crio-cf4366eab9a08e12a5593e952fa552a662878dea81a174ba3e0b2bb58875bf05 WatchSource:0}: Error finding container cf4366eab9a08e12a5593e952fa552a662878dea81a174ba3e0b2bb58875bf05: Status 404 returned error can't find the container with id cf4366eab9a08e12a5593e952fa552a662878dea81a174ba3e0b2bb58875bf05 Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.372061 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.395020 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kkj24"] Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.409358 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.410139 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:53.910124517 +0000 UTC m=+98.016535400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.487682 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:53 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:53 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:53 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.487728 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.513386 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.513761 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.013731918 +0000 UTC m=+98.120142801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.561640 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qrllc" Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.616401 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.616734 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.116724204 +0000 UTC m=+98.223135077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.717957 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.718154 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.218121 +0000 UTC m=+98.324531903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.718187 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.718568 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.218554681 +0000 UTC m=+98.324965564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.729000 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tzk59"] Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.811148 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerStarted","Data":"9849009f33e28e82aed7b22c035301860563d53bbde2ab038ccc58ab1bf958cc"} Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.814314 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" event={"ID":"4db3c52f-8fff-4625-ac11-c931baac1233","Type":"ContainerStarted","Data":"b746ff7ef6e8187a61aff2de771558fbf9ab15750f562e4c590db8901a973498"} Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.819839 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.820382 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.320363757 +0000 UTC m=+98.426774640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.821076 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerStarted","Data":"30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4"} Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.821126 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerStarted","Data":"d2d13b545b44e820836dfa0d6e88ac777a7259a59cbb66ff5a313b9e2a3c3fb4"} Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.830178 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerStarted","Data":"0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2"} Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.830272 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerStarted","Data":"cf4366eab9a08e12a5593e952fa552a662878dea81a174ba3e0b2bb58875bf05"} Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.863798 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxfsf"] Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.922520 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:53 crc kubenswrapper[4710]: E1002 12:50:53.923447 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.423408704 +0000 UTC m=+98.529819597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:53 crc kubenswrapper[4710]: I1002 12:50:53.939662 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.025227 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.025841 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.525822815 +0000 UTC m=+98.632233698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.127863 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.128437 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.628420951 +0000 UTC m=+98.734831834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.229847 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.230058 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.730033102 +0000 UTC m=+98.836443995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.230355 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.230712 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.730704029 +0000 UTC m=+98.837114912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.331117 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.331310 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.831284324 +0000 UTC m=+98.937695207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.331565 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.331894 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.831882769 +0000 UTC m=+98.938293652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.416947 4710 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.432301 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.432586 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.932560067 +0000 UTC m=+99.038970950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.432896 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.433356 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:54.933348927 +0000 UTC m=+99.039759810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.441955 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vwzwf"] Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.442912 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.444698 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.457702 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vwzwf"] Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.478570 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:54 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:54 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:54 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.478621 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.509358 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.509339945 podStartE2EDuration="1.509339945s" podCreationTimestamp="2025-10-02 12:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:54.485362503 +0000 UTC m=+98.591773406" watchObservedRunningTime="2025-10-02 12:50:54.509339945 +0000 UTC m=+98.615750828" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.534034 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.534281 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qks7\" (UniqueName: \"kubernetes.io/projected/0ec33244-0383-45c4-aecc-17d4c3f4435c-kube-api-access-5qks7\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.534362 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-utilities\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.534398 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-catalog-content\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.534578 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.034562998 +0000 UTC m=+99.140973881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.636221 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.636614 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.13660126 +0000 UTC m=+99.243012143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.636918 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qks7\" (UniqueName: \"kubernetes.io/projected/0ec33244-0383-45c4-aecc-17d4c3f4435c-kube-api-access-5qks7\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.637015 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-utilities\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.637372 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-catalog-content\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.637492 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-utilities\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.637696 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-catalog-content\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.669567 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qks7\" (UniqueName: \"kubernetes.io/projected/0ec33244-0383-45c4-aecc-17d4c3f4435c-kube-api-access-5qks7\") pod \"redhat-marketplace-vwzwf\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.738646 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.738809 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.238782595 +0000 UTC m=+99.345193468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.738961 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.739247 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.239238997 +0000 UTC m=+99.345649880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.761671 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.824128 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pxb8g"] Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.825083 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.838946 4710 generic.go:334] "Generic (PLEG): container finished" podID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerID="60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3" exitCode=0 Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.839448 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.839543 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.339525095 +0000 UTC m=+99.445935978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.839733 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.840193 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.340174271 +0000 UTC m=+99.446585244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.840814 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxfsf" event={"ID":"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5","Type":"ContainerDied","Data":"60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.840870 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxfsf" event={"ID":"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5","Type":"ContainerStarted","Data":"999517f3675887da960770e3751c143d6b3d884945f2398d8f0247ac5017b5f9"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.842711 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.850201 4710 generic.go:334] "Generic (PLEG): container finished" podID="8527e0f4-d514-45e1-89db-2d07c359387d" containerID="30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4" exitCode=0 Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.850464 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerDied","Data":"30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.853509 4710 generic.go:334] "Generic (PLEG): container finished" podID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerID="0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2" exitCode=0 Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.853558 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerDied","Data":"0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.859286 4710 generic.go:334] "Generic (PLEG): container finished" podID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerID="3c545cddabf9b8e898b9f6242e87b51a66564761fbaf3f7eb3f5d351a6f0fdf1" exitCode=0 Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.859353 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerDied","Data":"3c545cddabf9b8e898b9f6242e87b51a66564761fbaf3f7eb3f5d351a6f0fdf1"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.864135 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" event={"ID":"4db3c52f-8fff-4625-ac11-c931baac1233","Type":"ContainerStarted","Data":"a2f6fb07dd805bacfb6968456f670a8e07cd203345a30d4b24185b07eeee8516"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.864191 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" event={"ID":"4db3c52f-8fff-4625-ac11-c931baac1233","Type":"ContainerStarted","Data":"3b6c6bc7ed0f15f4dc1ac7586b9ae2bd920eff64708771b58660c69c9560a6bd"} Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.884821 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxb8g"] Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.946274 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.946730 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.446699896 +0000 UTC m=+99.553110779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.947092 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rhbn\" (UniqueName: \"kubernetes.io/projected/0ab22687-b3c6-4efa-8926-bde00f550f82-kube-api-access-6rhbn\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.947145 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.947276 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-utilities\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.947345 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-catalog-content\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:54 crc kubenswrapper[4710]: E1002 12:50:54.948112 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 12:50:55.448090871 +0000 UTC m=+99.554501754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cx6br" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 12:50:54 crc kubenswrapper[4710]: I1002 12:50:54.963462 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-xmcpk" podStartSLOduration=10.963447106 podStartE2EDuration="10.963447106s" podCreationTimestamp="2025-10-02 12:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:54.962255656 +0000 UTC m=+99.068666539" watchObservedRunningTime="2025-10-02 12:50:54.963447106 +0000 UTC m=+99.069857989" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.011602 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vwzwf"] Oct 02 12:50:55 crc kubenswrapper[4710]: W1002 12:50:55.022112 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ec33244_0383_45c4_aecc_17d4c3f4435c.slice/crio-a4a3425314af0b543b1fa2bc803bc70810960c74a1fc25bea6aeda693c8542d6 WatchSource:0}: Error finding container a4a3425314af0b543b1fa2bc803bc70810960c74a1fc25bea6aeda693c8542d6: Status 404 returned error can't find the container with id a4a3425314af0b543b1fa2bc803bc70810960c74a1fc25bea6aeda693c8542d6 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.033777 4710 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T12:50:54.416993736Z","Handler":null,"Name":""} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.036155 4710 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.036189 4710 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.048961 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.049242 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rhbn\" (UniqueName: \"kubernetes.io/projected/0ab22687-b3c6-4efa-8926-bde00f550f82-kube-api-access-6rhbn\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.049279 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-utilities\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.049312 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-catalog-content\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.049825 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-catalog-content\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.050355 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-utilities\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.054830 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.072861 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rhbn\" (UniqueName: \"kubernetes.io/projected/0ab22687-b3c6-4efa-8926-bde00f550f82-kube-api-access-6rhbn\") pod \"redhat-marketplace-pxb8g\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.144005 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.150609 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.156729 4710 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.156791 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.218763 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cx6br\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.315063 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.315954 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.318131 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.318634 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.322341 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.327392 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxb8g"] Oct 02 12:50:55 crc kubenswrapper[4710]: W1002 12:50:55.453352 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ab22687_b3c6_4efa_8926_bde00f550f82.slice/crio-de7067366ae4d8e2608ad05a57cdcc448f01178a976a08c8e08e37e6d5453146 WatchSource:0}: Error finding container de7067366ae4d8e2608ad05a57cdcc448f01178a976a08c8e08e37e6d5453146: Status 404 returned error can't find the container with id de7067366ae4d8e2608ad05a57cdcc448f01178a976a08c8e08e37e6d5453146 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.454954 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e18ebde3-f14d-43df-b06c-5a2be890d56b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.455077 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e18ebde3-f14d-43df-b06c-5a2be890d56b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.478158 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:55 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:55 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:55 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.478258 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.494784 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.556538 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e18ebde3-f14d-43df-b06c-5a2be890d56b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.557007 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e18ebde3-f14d-43df-b06c-5a2be890d56b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.556785 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e18ebde3-f14d-43df-b06c-5a2be890d56b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.575321 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e18ebde3-f14d-43df-b06c-5a2be890d56b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.662299 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rzt2p"] Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.664625 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.669172 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.683650 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzt2p"] Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.756310 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.761054 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9nz\" (UniqueName: \"kubernetes.io/projected/07062fef-42ef-410d-adc5-0ce1063b6f26-kube-api-access-sf9nz\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.761144 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-utilities\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.761190 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-catalog-content\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.771182 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cx6br"] Oct 02 12:50:55 crc kubenswrapper[4710]: W1002 12:50:55.784101 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb597145_79db_4f18_ae0c_c64f35af2fcc.slice/crio-0a2ad9d40366a0005d408a7f3616925d8c6ff199fb192e9a35ca1c4bd79d2f77 WatchSource:0}: Error finding container 0a2ad9d40366a0005d408a7f3616925d8c6ff199fb192e9a35ca1c4bd79d2f77: Status 404 returned error can't find the container with id 0a2ad9d40366a0005d408a7f3616925d8c6ff199fb192e9a35ca1c4bd79d2f77 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.862706 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9nz\" (UniqueName: \"kubernetes.io/projected/07062fef-42ef-410d-adc5-0ce1063b6f26-kube-api-access-sf9nz\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.862799 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-utilities\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.862863 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-catalog-content\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.863279 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-catalog-content\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.863336 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-utilities\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.871404 4710 generic.go:334] "Generic (PLEG): container finished" podID="e8205f2e-3067-4ab0-9e30-43d99dfd53eb" containerID="615d81488b245ad7e8fb43787023141b4ba3fc1d197076bceafe1fe7a4195f68" exitCode=0 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.871461 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" event={"ID":"e8205f2e-3067-4ab0-9e30-43d99dfd53eb","Type":"ContainerDied","Data":"615d81488b245ad7e8fb43787023141b4ba3fc1d197076bceafe1fe7a4195f68"} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.876050 4710 generic.go:334] "Generic (PLEG): container finished" podID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerID="845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169" exitCode=0 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.876324 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxb8g" event={"ID":"0ab22687-b3c6-4efa-8926-bde00f550f82","Type":"ContainerDied","Data":"845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169"} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.876374 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxb8g" event={"ID":"0ab22687-b3c6-4efa-8926-bde00f550f82","Type":"ContainerStarted","Data":"de7067366ae4d8e2608ad05a57cdcc448f01178a976a08c8e08e37e6d5453146"} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.879033 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9nz\" (UniqueName: \"kubernetes.io/projected/07062fef-42ef-410d-adc5-0ce1063b6f26-kube-api-access-sf9nz\") pod \"redhat-operators-rzt2p\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.879053 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" event={"ID":"db597145-79db-4f18-ae0c-c64f35af2fcc","Type":"ContainerStarted","Data":"0a2ad9d40366a0005d408a7f3616925d8c6ff199fb192e9a35ca1c4bd79d2f77"} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.882619 4710 generic.go:334] "Generic (PLEG): container finished" podID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerID="6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7" exitCode=0 Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.882773 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vwzwf" event={"ID":"0ec33244-0383-45c4-aecc-17d4c3f4435c","Type":"ContainerDied","Data":"6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7"} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.882832 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vwzwf" event={"ID":"0ec33244-0383-45c4-aecc-17d4c3f4435c","Type":"ContainerStarted","Data":"a4a3425314af0b543b1fa2bc803bc70810960c74a1fc25bea6aeda693c8542d6"} Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.963800 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.968304 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d77362b1-cb9c-498f-8e21-9114ce2de0d3-metrics-certs\") pod \"network-metrics-daemon-zqx98\" (UID: \"d77362b1-cb9c-498f-8e21-9114ce2de0d3\") " pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:55 crc kubenswrapper[4710]: I1002 12:50:55.999354 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.030073 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rsbdg"] Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.031154 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.040430 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rsbdg"] Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.116508 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.116604 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zqx98" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.168629 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhsbt\" (UniqueName: \"kubernetes.io/projected/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-kube-api-access-vhsbt\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.168780 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-utilities\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.168811 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-catalog-content\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.270296 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-utilities\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.270335 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-catalog-content\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.270367 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhsbt\" (UniqueName: \"kubernetes.io/projected/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-kube-api-access-vhsbt\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.271243 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-utilities\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.271589 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-catalog-content\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.286997 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhsbt\" (UniqueName: \"kubernetes.io/projected/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-kube-api-access-vhsbt\") pod \"redhat-operators-rsbdg\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.330421 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zqx98"] Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.354654 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.375202 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzt2p"] Oct 02 12:50:56 crc kubenswrapper[4710]: W1002 12:50:56.398717 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07062fef_42ef_410d_adc5_0ce1063b6f26.slice/crio-31730b266111f49d9b701fb3253f0e92134b38449eb896f8c9a18330d900dc78 WatchSource:0}: Error finding container 31730b266111f49d9b701fb3253f0e92134b38449eb896f8c9a18330d900dc78: Status 404 returned error can't find the container with id 31730b266111f49d9b701fb3253f0e92134b38449eb896f8c9a18330d900dc78 Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.479027 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:56 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:56 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:56 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.479087 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.583102 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rsbdg"] Oct 02 12:50:56 crc kubenswrapper[4710]: W1002 12:50:56.588009 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf1a2d7d_ccb8_4532_b6b1_2da89b14a32d.slice/crio-e6171239d42a0497fd701da9f814676f5562f5390c8ab29f49848ab4b155f466 WatchSource:0}: Error finding container e6171239d42a0497fd701da9f814676f5562f5390c8ab29f49848ab4b155f466: Status 404 returned error can't find the container with id e6171239d42a0497fd701da9f814676f5562f5390c8ab29f49848ab4b155f466 Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.828509 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.836864 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-c5sq4" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.894142 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e18ebde3-f14d-43df-b06c-5a2be890d56b","Type":"ContainerStarted","Data":"d845b3f8d9e7eadfa1baa3c16348fa503d855c616ebef47272b081fcf08f242c"} Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.894191 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e18ebde3-f14d-43df-b06c-5a2be890d56b","Type":"ContainerStarted","Data":"e37048ca7ef01e4dfe5dea86330f630ecae37e2a9d3324446b12280fc7dbf642"} Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.895633 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" event={"ID":"db597145-79db-4f18-ae0c-c64f35af2fcc","Type":"ContainerStarted","Data":"478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019"} Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.895736 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.896698 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerStarted","Data":"e6171239d42a0497fd701da9f814676f5562f5390c8ab29f49848ab4b155f466"} Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.897515 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zqx98" event={"ID":"d77362b1-cb9c-498f-8e21-9114ce2de0d3","Type":"ContainerStarted","Data":"bfe8a748b848e1527c9fc32a3bda351ead687a5b34cb91929617b4a407fc1814"} Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.898816 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerStarted","Data":"31730b266111f49d9b701fb3253f0e92134b38449eb896f8c9a18330d900dc78"} Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.941979 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" podStartSLOduration=79.94195683 podStartE2EDuration="1m19.94195683s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:56.938264387 +0000 UTC m=+101.044675280" watchObservedRunningTime="2025-10-02 12:50:56.94195683 +0000 UTC m=+101.048367713" Oct 02 12:50:56 crc kubenswrapper[4710]: I1002 12:50:56.996647 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.155536 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.155910 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.157479 4710 patch_prober.go:28] interesting pod/console-f9d7485db-mrtmb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.157528 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mrtmb" podUID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.222555 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.263575 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.296308 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-config-volume\") pod \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.296360 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-secret-volume\") pod \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.296385 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8746\" (UniqueName: \"kubernetes.io/projected/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-kube-api-access-b8746\") pod \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\" (UID: \"e8205f2e-3067-4ab0-9e30-43d99dfd53eb\") " Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.297126 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "e8205f2e-3067-4ab0-9e30-43d99dfd53eb" (UID: "e8205f2e-3067-4ab0-9e30-43d99dfd53eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.312267 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-kube-api-access-b8746" (OuterVolumeSpecName: "kube-api-access-b8746") pod "e8205f2e-3067-4ab0-9e30-43d99dfd53eb" (UID: "e8205f2e-3067-4ab0-9e30-43d99dfd53eb"). InnerVolumeSpecName "kube-api-access-b8746". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.316355 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e8205f2e-3067-4ab0-9e30-43d99dfd53eb" (UID: "e8205f2e-3067-4ab0-9e30-43d99dfd53eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.374822 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.374876 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.375549 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.375611 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.397688 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.397729 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8746\" (UniqueName: \"kubernetes.io/projected/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-kube-api-access-b8746\") on node \"crc\" DevicePath \"\"" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.397756 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8205f2e-3067-4ab0-9e30-43d99dfd53eb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.475264 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.478800 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:57 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:57 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:57 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.478848 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.956214 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zqx98" event={"ID":"d77362b1-cb9c-498f-8e21-9114ce2de0d3","Type":"ContainerStarted","Data":"56bbef6e049c5e35de805d9a1f43dbbb8591612b8d2595bc88cc73c0d08c1cfb"} Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.956494 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zqx98" event={"ID":"d77362b1-cb9c-498f-8e21-9114ce2de0d3","Type":"ContainerStarted","Data":"302585a6bb9af969444bbf518e0750ed4c09acf025cd6c43bfed9ecd74a78872"} Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.968930 4710 generic.go:334] "Generic (PLEG): container finished" podID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerID="fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606" exitCode=0 Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.968998 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerDied","Data":"fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606"} Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.974819 4710 generic.go:334] "Generic (PLEG): container finished" podID="e18ebde3-f14d-43df-b06c-5a2be890d56b" containerID="d845b3f8d9e7eadfa1baa3c16348fa503d855c616ebef47272b081fcf08f242c" exitCode=0 Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.974950 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e18ebde3-f14d-43df-b06c-5a2be890d56b","Type":"ContainerDied","Data":"d845b3f8d9e7eadfa1baa3c16348fa503d855c616ebef47272b081fcf08f242c"} Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.979042 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.979161 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz" event={"ID":"e8205f2e-3067-4ab0-9e30-43d99dfd53eb","Type":"ContainerDied","Data":"2a31f35a0c7ea30b986761fb35e2f9a1337cbe48a6da9fe6bd96035bc3c7c59c"} Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.979186 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a31f35a0c7ea30b986761fb35e2f9a1337cbe48a6da9fe6bd96035bc3c7c59c" Oct 02 12:50:57 crc kubenswrapper[4710]: I1002 12:50:57.986250 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zqx98" podStartSLOduration=80.986197528 podStartE2EDuration="1m20.986197528s" podCreationTimestamp="2025-10-02 12:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:50:57.971832247 +0000 UTC m=+102.078243140" watchObservedRunningTime="2025-10-02 12:50:57.986197528 +0000 UTC m=+102.092608411" Oct 02 12:50:58 crc kubenswrapper[4710]: I1002 12:50:58.018355 4710 generic.go:334] "Generic (PLEG): container finished" podID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerID="2312b8d9a45c3a376732fe75660fa57b8acf33be62463e0baa0e7c78dde6315c" exitCode=0 Oct 02 12:50:58 crc kubenswrapper[4710]: I1002 12:50:58.019455 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerDied","Data":"2312b8d9a45c3a376732fe75660fa57b8acf33be62463e0baa0e7c78dde6315c"} Oct 02 12:50:58 crc kubenswrapper[4710]: I1002 12:50:58.261880 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:50:58 crc kubenswrapper[4710]: I1002 12:50:58.479287 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:58 crc kubenswrapper[4710]: [-]has-synced failed: reason withheld Oct 02 12:50:58 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:58 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:58 crc kubenswrapper[4710]: I1002 12:50:58.479336 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.368558 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.442979 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e18ebde3-f14d-43df-b06c-5a2be890d56b-kubelet-dir\") pod \"e18ebde3-f14d-43df-b06c-5a2be890d56b\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.443090 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e18ebde3-f14d-43df-b06c-5a2be890d56b-kube-api-access\") pod \"e18ebde3-f14d-43df-b06c-5a2be890d56b\" (UID: \"e18ebde3-f14d-43df-b06c-5a2be890d56b\") " Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.443092 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e18ebde3-f14d-43df-b06c-5a2be890d56b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e18ebde3-f14d-43df-b06c-5a2be890d56b" (UID: "e18ebde3-f14d-43df-b06c-5a2be890d56b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.443367 4710 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e18ebde3-f14d-43df-b06c-5a2be890d56b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.466953 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18ebde3-f14d-43df-b06c-5a2be890d56b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e18ebde3-f14d-43df-b06c-5a2be890d56b" (UID: "e18ebde3-f14d-43df-b06c-5a2be890d56b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.479676 4710 patch_prober.go:28] interesting pod/router-default-5444994796-j786v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 12:50:59 crc kubenswrapper[4710]: [+]has-synced ok Oct 02 12:50:59 crc kubenswrapper[4710]: [+]process-running ok Oct 02 12:50:59 crc kubenswrapper[4710]: healthz check failed Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.479733 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j786v" podUID="c8b54dd0-afe8-4f80-ab3d-db6040735d5a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 12:50:59 crc kubenswrapper[4710]: I1002 12:50:59.544525 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e18ebde3-f14d-43df-b06c-5a2be890d56b-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.044143 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 12:51:00 crc kubenswrapper[4710]: E1002 12:51:00.044382 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18ebde3-f14d-43df-b06c-5a2be890d56b" containerName="pruner" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.044393 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18ebde3-f14d-43df-b06c-5a2be890d56b" containerName="pruner" Oct 02 12:51:00 crc kubenswrapper[4710]: E1002 12:51:00.044412 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8205f2e-3067-4ab0-9e30-43d99dfd53eb" containerName="collect-profiles" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.044418 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8205f2e-3067-4ab0-9e30-43d99dfd53eb" containerName="collect-profiles" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.044508 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18ebde3-f14d-43df-b06c-5a2be890d56b" containerName="pruner" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.044520 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8205f2e-3067-4ab0-9e30-43d99dfd53eb" containerName="collect-profiles" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.044914 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.049161 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.049572 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.051300 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.064281 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e18ebde3-f14d-43df-b06c-5a2be890d56b","Type":"ContainerDied","Data":"e37048ca7ef01e4dfe5dea86330f630ecae37e2a9d3324446b12280fc7dbf642"} Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.064335 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e37048ca7ef01e4dfe5dea86330f630ecae37e2a9d3324446b12280fc7dbf642" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.064433 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.163218 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8612401-3e78-458a-ad2d-746b2149d7c4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.163551 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8612401-3e78-458a-ad2d-746b2149d7c4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.264275 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8612401-3e78-458a-ad2d-746b2149d7c4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.264338 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8612401-3e78-458a-ad2d-746b2149d7c4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.264378 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8612401-3e78-458a-ad2d-746b2149d7c4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.320167 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8612401-3e78-458a-ad2d-746b2149d7c4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.398150 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.480456 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.482651 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-j786v" Oct 02 12:51:00 crc kubenswrapper[4710]: I1002 12:51:00.887591 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 12:51:01 crc kubenswrapper[4710]: I1002 12:51:01.081960 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a8612401-3e78-458a-ad2d-746b2149d7c4","Type":"ContainerStarted","Data":"27603cf21768ad928d6546bcda773e819c7a154b8b8870fda89399d50451d3ac"} Oct 02 12:51:02 crc kubenswrapper[4710]: I1002 12:51:02.665908 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-xt66p" Oct 02 12:51:03 crc kubenswrapper[4710]: I1002 12:51:03.104259 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a8612401-3e78-458a-ad2d-746b2149d7c4","Type":"ContainerStarted","Data":"74e25f198504199531e9578576498997a6158def6e7870aa5ed2545164250a05"} Oct 02 12:51:03 crc kubenswrapper[4710]: I1002 12:51:03.120477 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.120449992 podStartE2EDuration="3.120449992s" podCreationTimestamp="2025-10-02 12:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:51:03.117058947 +0000 UTC m=+107.223469840" watchObservedRunningTime="2025-10-02 12:51:03.120449992 +0000 UTC m=+107.226860875" Oct 02 12:51:04 crc kubenswrapper[4710]: I1002 12:51:04.113740 4710 generic.go:334] "Generic (PLEG): container finished" podID="a8612401-3e78-458a-ad2d-746b2149d7c4" containerID="74e25f198504199531e9578576498997a6158def6e7870aa5ed2545164250a05" exitCode=0 Oct 02 12:51:04 crc kubenswrapper[4710]: I1002 12:51:04.113798 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a8612401-3e78-458a-ad2d-746b2149d7c4","Type":"ContainerDied","Data":"74e25f198504199531e9578576498997a6158def6e7870aa5ed2545164250a05"} Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.629954 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.770067 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8612401-3e78-458a-ad2d-746b2149d7c4-kubelet-dir\") pod \"a8612401-3e78-458a-ad2d-746b2149d7c4\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.770206 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8612401-3e78-458a-ad2d-746b2149d7c4-kube-api-access\") pod \"a8612401-3e78-458a-ad2d-746b2149d7c4\" (UID: \"a8612401-3e78-458a-ad2d-746b2149d7c4\") " Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.770222 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8612401-3e78-458a-ad2d-746b2149d7c4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a8612401-3e78-458a-ad2d-746b2149d7c4" (UID: "a8612401-3e78-458a-ad2d-746b2149d7c4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.770559 4710 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8612401-3e78-458a-ad2d-746b2149d7c4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.777349 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8612401-3e78-458a-ad2d-746b2149d7c4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a8612401-3e78-458a-ad2d-746b2149d7c4" (UID: "a8612401-3e78-458a-ad2d-746b2149d7c4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:51:05 crc kubenswrapper[4710]: I1002 12:51:05.871290 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8612401-3e78-458a-ad2d-746b2149d7c4-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 12:51:06 crc kubenswrapper[4710]: I1002 12:51:06.132818 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 12:51:06 crc kubenswrapper[4710]: I1002 12:51:06.134468 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a8612401-3e78-458a-ad2d-746b2149d7c4","Type":"ContainerDied","Data":"27603cf21768ad928d6546bcda773e819c7a154b8b8870fda89399d50451d3ac"} Oct 02 12:51:06 crc kubenswrapper[4710]: I1002 12:51:06.134621 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27603cf21768ad928d6546bcda773e819c7a154b8b8870fda89399d50451d3ac" Oct 02 12:51:07 crc kubenswrapper[4710]: I1002 12:51:07.159805 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:51:07 crc kubenswrapper[4710]: I1002 12:51:07.163599 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 12:51:07 crc kubenswrapper[4710]: I1002 12:51:07.374230 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:07 crc kubenswrapper[4710]: I1002 12:51:07.374322 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:07 crc kubenswrapper[4710]: I1002 12:51:07.374443 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:07 crc kubenswrapper[4710]: I1002 12:51:07.374531 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:15 crc kubenswrapper[4710]: I1002 12:51:15.505957 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.374619 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.374629 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.375597 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.375649 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.375734 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.376348 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.376398 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.376857 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"ded0e24b3cfd20b23355b16d1e83f7e62f1cebb0eb615beed9f6f3c6b8e311e8"} pod="openshift-console/downloads-7954f5f757-m5hbd" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 02 12:51:17 crc kubenswrapper[4710]: I1002 12:51:17.377000 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" containerID="cri-o://ded0e24b3cfd20b23355b16d1e83f7e62f1cebb0eb615beed9f6f3c6b8e311e8" gracePeriod=2 Oct 02 12:51:19 crc kubenswrapper[4710]: I1002 12:51:19.224862 4710 generic.go:334] "Generic (PLEG): container finished" podID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerID="ded0e24b3cfd20b23355b16d1e83f7e62f1cebb0eb615beed9f6f3c6b8e311e8" exitCode=0 Oct 02 12:51:19 crc kubenswrapper[4710]: I1002 12:51:19.224991 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-m5hbd" event={"ID":"c68e372e-5db7-475e-bacb-5452d32cbdfc","Type":"ContainerDied","Data":"ded0e24b3cfd20b23355b16d1e83f7e62f1cebb0eb615beed9f6f3c6b8e311e8"} Oct 02 12:51:27 crc kubenswrapper[4710]: I1002 12:51:27.374697 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:27 crc kubenswrapper[4710]: I1002 12:51:27.375142 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:27 crc kubenswrapper[4710]: I1002 12:51:27.797317 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qpnbj" Oct 02 12:51:29 crc kubenswrapper[4710]: E1002 12:51:29.908677 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 12:51:29 crc kubenswrapper[4710]: E1002 12:51:29.909238 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6rhbn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pxb8g_openshift-marketplace(0ab22687-b3c6-4efa-8926-bde00f550f82): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:29 crc kubenswrapper[4710]: E1002 12:51:29.910457 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pxb8g" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" Oct 02 12:51:30 crc kubenswrapper[4710]: E1002 12:51:30.615273 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pxb8g" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" Oct 02 12:51:32 crc kubenswrapper[4710]: E1002 12:51:32.110005 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 12:51:32 crc kubenswrapper[4710]: E1002 12:51:32.110492 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qks7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vwzwf_openshift-marketplace(0ec33244-0383-45c4-aecc-17d4c3f4435c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:32 crc kubenswrapper[4710]: E1002 12:51:32.113535 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vwzwf" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" Oct 02 12:51:35 crc kubenswrapper[4710]: E1002 12:51:35.854739 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vwzwf" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" Oct 02 12:51:37 crc kubenswrapper[4710]: I1002 12:51:37.373739 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:37 crc kubenswrapper[4710]: I1002 12:51:37.373817 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.877270 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.877737 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.877928 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.878020 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.880479 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.880780 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.882109 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.890812 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.898385 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.903309 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.904772 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:51:44 crc kubenswrapper[4710]: I1002 12:51:44.987603 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 12:51:45 crc kubenswrapper[4710]: I1002 12:51:45.001089 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:51:45 crc kubenswrapper[4710]: I1002 12:51:45.344985 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:51:45 crc kubenswrapper[4710]: I1002 12:51:45.595083 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 12:51:47 crc kubenswrapper[4710]: I1002 12:51:47.376692 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:47 crc kubenswrapper[4710]: I1002 12:51:47.377146 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:51 crc kubenswrapper[4710]: E1002 12:51:51.869120 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 12:51:51 crc kubenswrapper[4710]: E1002 12:51:51.869835 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vhsbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rsbdg_openshift-marketplace(af1a2d7d-ccb8-4532-b6b1-2da89b14a32d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:51 crc kubenswrapper[4710]: E1002 12:51:51.871031 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rsbdg" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" Oct 02 12:51:52 crc kubenswrapper[4710]: I1002 12:51:52.530244 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:51:52 crc kubenswrapper[4710]: I1002 12:51:52.530581 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:51:53 crc kubenswrapper[4710]: E1002 12:51:53.828591 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rsbdg" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" Oct 02 12:51:56 crc kubenswrapper[4710]: E1002 12:51:56.867376 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 12:51:56 crc kubenswrapper[4710]: E1002 12:51:56.867546 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwr5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tzk59_openshift-marketplace(2ef28114-3b05-44a8-b1e5-5bbbda54d657): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:56 crc kubenswrapper[4710]: E1002 12:51:56.868825 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tzk59" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" Oct 02 12:51:57 crc kubenswrapper[4710]: I1002 12:51:57.374392 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:57 crc kubenswrapper[4710]: I1002 12:51:57.374918 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.183285 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tzk59" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.261113 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.261287 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bhsw8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7tgfs_openshift-marketplace(567598d7-6c35-46b0-bd80-5c04b4f7e024): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.262461 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7tgfs" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.269972 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.270145 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6hpw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kxfsf_openshift-marketplace(1535dfb1-9951-4d3e-ace9-2c70a4d0abb5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.271246 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kxfsf" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.348265 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.348704 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sf9nz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rzt2p_openshift-marketplace(07062fef-42ef-410d-adc5-0ce1063b6f26): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.349811 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rzt2p" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.365589 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.365772 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8pvsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kkj24_openshift-marketplace(8527e0f4-d514-45e1-89db-2d07c359387d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.366988 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kkj24" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" Oct 02 12:51:58 crc kubenswrapper[4710]: I1002 12:51:58.485060 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"eae9f8946be4fb3b1f942be3bd4f138b4179705e1cc35e3dbd536c7ace3565e9"} Oct 02 12:51:58 crc kubenswrapper[4710]: I1002 12:51:58.498456 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-m5hbd" event={"ID":"c68e372e-5db7-475e-bacb-5452d32cbdfc","Type":"ContainerStarted","Data":"ec9b64d5ec3affe3604b5db27d74f60265c466af926bf4f5d1a8655034725447"} Oct 02 12:51:58 crc kubenswrapper[4710]: I1002 12:51:58.504654 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:58 crc kubenswrapper[4710]: I1002 12:51:58.504709 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.505270 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7tgfs" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.505627 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kkj24" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.505709 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kxfsf" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" Oct 02 12:51:58 crc kubenswrapper[4710]: E1002 12:51:58.505774 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rzt2p" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.505314 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"781c03464ef25862e003420ed272206b80579377bc013b8e0ccc53a23ac543ec"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.505658 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"dd01ea8b0b7d6c4d1d51eb6e11cb4c2a2d9f297bdc7f8265d6b95cafe0c10613"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.507755 4710 generic.go:334] "Generic (PLEG): container finished" podID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerID="0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61" exitCode=0 Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.507819 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vwzwf" event={"ID":"0ec33244-0383-45c4-aecc-17d4c3f4435c","Type":"ContainerDied","Data":"0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.511305 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9a0bf7d2a082b89993b13c6710850d34ba0df26d810dbc0141fa0d747070fba7"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.511384 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"99d1c834a18f2162c3b03ac331232ecc90fc6bfe2c3da4288db05f597c488b4c"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.514886 4710 generic.go:334] "Generic (PLEG): container finished" podID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerID="9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b" exitCode=0 Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.514977 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxb8g" event={"ID":"0ab22687-b3c6-4efa-8926-bde00f550f82","Type":"ContainerDied","Data":"9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.517395 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ff6c640b09f16034996d929a3a8b383fd3bedd3160fa2f2394927ee00097569f"} Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.517598 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.517673 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.518015 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:51:59 crc kubenswrapper[4710]: I1002 12:51:59.518070 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:52:00 crc kubenswrapper[4710]: I1002 12:52:00.526007 4710 patch_prober.go:28] interesting pod/downloads-7954f5f757-m5hbd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 12:52:00 crc kubenswrapper[4710]: I1002 12:52:00.526505 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m5hbd" podUID="c68e372e-5db7-475e-bacb-5452d32cbdfc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 12:52:01 crc kubenswrapper[4710]: I1002 12:52:01.531035 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vwzwf" event={"ID":"0ec33244-0383-45c4-aecc-17d4c3f4435c","Type":"ContainerStarted","Data":"ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2"} Oct 02 12:52:01 crc kubenswrapper[4710]: I1002 12:52:01.534365 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxb8g" event={"ID":"0ab22687-b3c6-4efa-8926-bde00f550f82","Type":"ContainerStarted","Data":"d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530"} Oct 02 12:52:01 crc kubenswrapper[4710]: I1002 12:52:01.569941 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vwzwf" podStartSLOduration=2.708683724 podStartE2EDuration="1m7.569905162s" podCreationTimestamp="2025-10-02 12:50:54 +0000 UTC" firstStartedPulling="2025-10-02 12:50:55.888367507 +0000 UTC m=+99.994778390" lastFinishedPulling="2025-10-02 12:52:00.749588935 +0000 UTC m=+164.855999828" observedRunningTime="2025-10-02 12:52:01.565335187 +0000 UTC m=+165.671746070" watchObservedRunningTime="2025-10-02 12:52:01.569905162 +0000 UTC m=+165.676316085" Oct 02 12:52:01 crc kubenswrapper[4710]: I1002 12:52:01.590660 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pxb8g" podStartSLOduration=2.979626648 podStartE2EDuration="1m7.590635444s" podCreationTimestamp="2025-10-02 12:50:54 +0000 UTC" firstStartedPulling="2025-10-02 12:50:55.87970062 +0000 UTC m=+99.986111503" lastFinishedPulling="2025-10-02 12:52:00.490709416 +0000 UTC m=+164.597120299" observedRunningTime="2025-10-02 12:52:01.587023097 +0000 UTC m=+165.693434050" watchObservedRunningTime="2025-10-02 12:52:01.590635444 +0000 UTC m=+165.697046327" Oct 02 12:52:04 crc kubenswrapper[4710]: I1002 12:52:04.762324 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:52:04 crc kubenswrapper[4710]: I1002 12:52:04.762732 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:52:05 crc kubenswrapper[4710]: I1002 12:52:05.144813 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:52:05 crc kubenswrapper[4710]: I1002 12:52:05.145244 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:52:05 crc kubenswrapper[4710]: I1002 12:52:05.324523 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:52:05 crc kubenswrapper[4710]: I1002 12:52:05.329081 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:52:05 crc kubenswrapper[4710]: I1002 12:52:05.609546 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:52:05 crc kubenswrapper[4710]: I1002 12:52:05.638428 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:52:06 crc kubenswrapper[4710]: I1002 12:52:06.738167 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxb8g"] Oct 02 12:52:07 crc kubenswrapper[4710]: I1002 12:52:07.383326 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-m5hbd" Oct 02 12:52:07 crc kubenswrapper[4710]: I1002 12:52:07.570947 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pxb8g" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="registry-server" containerID="cri-o://d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530" gracePeriod=2 Oct 02 12:52:07 crc kubenswrapper[4710]: I1002 12:52:07.959173 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.109919 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-utilities\") pod \"0ab22687-b3c6-4efa-8926-bde00f550f82\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.110143 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rhbn\" (UniqueName: \"kubernetes.io/projected/0ab22687-b3c6-4efa-8926-bde00f550f82-kube-api-access-6rhbn\") pod \"0ab22687-b3c6-4efa-8926-bde00f550f82\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.110205 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-catalog-content\") pod \"0ab22687-b3c6-4efa-8926-bde00f550f82\" (UID: \"0ab22687-b3c6-4efa-8926-bde00f550f82\") " Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.111883 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-utilities" (OuterVolumeSpecName: "utilities") pod "0ab22687-b3c6-4efa-8926-bde00f550f82" (UID: "0ab22687-b3c6-4efa-8926-bde00f550f82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.116065 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab22687-b3c6-4efa-8926-bde00f550f82-kube-api-access-6rhbn" (OuterVolumeSpecName: "kube-api-access-6rhbn") pod "0ab22687-b3c6-4efa-8926-bde00f550f82" (UID: "0ab22687-b3c6-4efa-8926-bde00f550f82"). InnerVolumeSpecName "kube-api-access-6rhbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.148602 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ab22687-b3c6-4efa-8926-bde00f550f82" (UID: "0ab22687-b3c6-4efa-8926-bde00f550f82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.212043 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rhbn\" (UniqueName: \"kubernetes.io/projected/0ab22687-b3c6-4efa-8926-bde00f550f82-kube-api-access-6rhbn\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.212139 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.212169 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab22687-b3c6-4efa-8926-bde00f550f82-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.577960 4710 generic.go:334] "Generic (PLEG): container finished" podID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerID="d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530" exitCode=0 Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.578048 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxb8g" event={"ID":"0ab22687-b3c6-4efa-8926-bde00f550f82","Type":"ContainerDied","Data":"d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530"} Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.578098 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pxb8g" event={"ID":"0ab22687-b3c6-4efa-8926-bde00f550f82","Type":"ContainerDied","Data":"de7067366ae4d8e2608ad05a57cdcc448f01178a976a08c8e08e37e6d5453146"} Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.578129 4710 scope.go:117] "RemoveContainer" containerID="d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.578650 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pxb8g" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.600000 4710 scope.go:117] "RemoveContainer" containerID="9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.626649 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxb8g"] Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.630477 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pxb8g"] Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.638433 4710 scope.go:117] "RemoveContainer" containerID="845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.655976 4710 scope.go:117] "RemoveContainer" containerID="d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530" Oct 02 12:52:08 crc kubenswrapper[4710]: E1002 12:52:08.656710 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530\": container with ID starting with d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530 not found: ID does not exist" containerID="d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.656793 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530"} err="failed to get container status \"d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530\": rpc error: code = NotFound desc = could not find container \"d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530\": container with ID starting with d814031fa65768552990136aa19be0959cfb05fdc647eacb822e1b0d324d1530 not found: ID does not exist" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.656861 4710 scope.go:117] "RemoveContainer" containerID="9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b" Oct 02 12:52:08 crc kubenswrapper[4710]: E1002 12:52:08.659465 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b\": container with ID starting with 9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b not found: ID does not exist" containerID="9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.659516 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b"} err="failed to get container status \"9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b\": rpc error: code = NotFound desc = could not find container \"9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b\": container with ID starting with 9f7cc1a4acdc3e6601203a374af713dd9b503518d480a615420ab430ee5b4d7b not found: ID does not exist" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.659552 4710 scope.go:117] "RemoveContainer" containerID="845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169" Oct 02 12:52:08 crc kubenswrapper[4710]: E1002 12:52:08.659917 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169\": container with ID starting with 845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169 not found: ID does not exist" containerID="845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.659977 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169"} err="failed to get container status \"845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169\": rpc error: code = NotFound desc = could not find container \"845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169\": container with ID starting with 845b93f7e75ad37947ad5cdd78be2af72672f3465f05cdb74e5d0be881d58169 not found: ID does not exist" Oct 02 12:52:08 crc kubenswrapper[4710]: I1002 12:52:08.943851 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" path="/var/lib/kubelet/pods/0ab22687-b3c6-4efa-8926-bde00f550f82/volumes" Oct 02 12:52:10 crc kubenswrapper[4710]: I1002 12:52:10.607883 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerStarted","Data":"456cb6b188d03ff78d30a7d49370fc36603aec1ac8b7f28ece95bcb60e84d40e"} Oct 02 12:52:10 crc kubenswrapper[4710]: I1002 12:52:10.610950 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerStarted","Data":"6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667"} Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.617619 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerStarted","Data":"f2bf7535ab755fe5a77dd4d4766e605fc56405b5017885fe5b3096ae4f40d5d3"} Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.619829 4710 generic.go:334] "Generic (PLEG): container finished" podID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerID="456cb6b188d03ff78d30a7d49370fc36603aec1ac8b7f28ece95bcb60e84d40e" exitCode=0 Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.619903 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerDied","Data":"456cb6b188d03ff78d30a7d49370fc36603aec1ac8b7f28ece95bcb60e84d40e"} Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.621845 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerStarted","Data":"d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d"} Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.625005 4710 generic.go:334] "Generic (PLEG): container finished" podID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerID="9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02" exitCode=0 Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.625089 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxfsf" event={"ID":"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5","Type":"ContainerDied","Data":"9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02"} Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.629342 4710 generic.go:334] "Generic (PLEG): container finished" podID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerID="6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667" exitCode=0 Oct 02 12:52:11 crc kubenswrapper[4710]: I1002 12:52:11.629415 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerDied","Data":"6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667"} Oct 02 12:52:12 crc kubenswrapper[4710]: I1002 12:52:12.638637 4710 generic.go:334] "Generic (PLEG): container finished" podID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerID="f2bf7535ab755fe5a77dd4d4766e605fc56405b5017885fe5b3096ae4f40d5d3" exitCode=0 Oct 02 12:52:12 crc kubenswrapper[4710]: I1002 12:52:12.638720 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerDied","Data":"f2bf7535ab755fe5a77dd4d4766e605fc56405b5017885fe5b3096ae4f40d5d3"} Oct 02 12:52:12 crc kubenswrapper[4710]: I1002 12:52:12.641103 4710 generic.go:334] "Generic (PLEG): container finished" podID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerID="d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d" exitCode=0 Oct 02 12:52:12 crc kubenswrapper[4710]: I1002 12:52:12.641133 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerDied","Data":"d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d"} Oct 02 12:52:15 crc kubenswrapper[4710]: I1002 12:52:15.665107 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerStarted","Data":"4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc"} Oct 02 12:52:16 crc kubenswrapper[4710]: I1002 12:52:16.700690 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7tgfs" podStartSLOduration=4.480513801 podStartE2EDuration="1m24.700663383s" podCreationTimestamp="2025-10-02 12:50:52 +0000 UTC" firstStartedPulling="2025-10-02 12:50:54.854934412 +0000 UTC m=+98.961345295" lastFinishedPulling="2025-10-02 12:52:15.075083954 +0000 UTC m=+179.181494877" observedRunningTime="2025-10-02 12:52:16.694958025 +0000 UTC m=+180.801368918" watchObservedRunningTime="2025-10-02 12:52:16.700663383 +0000 UTC m=+180.807074306" Oct 02 12:52:18 crc kubenswrapper[4710]: I1002 12:52:18.685705 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerStarted","Data":"861126c2c34a093163cd527a0870fa75627b888715d8630a8a443ee6a6dff34d"} Oct 02 12:52:19 crc kubenswrapper[4710]: I1002 12:52:19.725407 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rsbdg" podStartSLOduration=4.380819475 podStartE2EDuration="1m23.725387239s" podCreationTimestamp="2025-10-02 12:50:56 +0000 UTC" firstStartedPulling="2025-10-02 12:50:58.032918411 +0000 UTC m=+102.139329294" lastFinishedPulling="2025-10-02 12:52:17.377486175 +0000 UTC m=+181.483897058" observedRunningTime="2025-10-02 12:52:19.722511104 +0000 UTC m=+183.828922017" watchObservedRunningTime="2025-10-02 12:52:19.725387239 +0000 UTC m=+183.831798132" Oct 02 12:52:22 crc kubenswrapper[4710]: I1002 12:52:22.530004 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:52:22 crc kubenswrapper[4710]: I1002 12:52:22.530345 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:52:22 crc kubenswrapper[4710]: I1002 12:52:22.951518 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:52:22 crc kubenswrapper[4710]: I1002 12:52:22.951609 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:52:23 crc kubenswrapper[4710]: I1002 12:52:23.035074 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:52:23 crc kubenswrapper[4710]: I1002 12:52:23.772121 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:52:26 crc kubenswrapper[4710]: I1002 12:52:26.355525 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:52:26 crc kubenswrapper[4710]: I1002 12:52:26.356037 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:52:26 crc kubenswrapper[4710]: I1002 12:52:26.406074 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:52:26 crc kubenswrapper[4710]: I1002 12:52:26.774583 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:52:27 crc kubenswrapper[4710]: I1002 12:52:27.105703 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rsbdg"] Oct 02 12:52:28 crc kubenswrapper[4710]: I1002 12:52:28.741065 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rsbdg" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="registry-server" containerID="cri-o://861126c2c34a093163cd527a0870fa75627b888715d8630a8a443ee6a6dff34d" gracePeriod=2 Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.750463 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxfsf" event={"ID":"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5","Type":"ContainerStarted","Data":"65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491"} Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.754947 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerStarted","Data":"d7535790593eca2803f42201a4983c8fd4d1bbed389f47e113911badfb9a9048"} Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.757306 4710 generic.go:334] "Generic (PLEG): container finished" podID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerID="861126c2c34a093163cd527a0870fa75627b888715d8630a8a443ee6a6dff34d" exitCode=0 Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.757348 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerDied","Data":"861126c2c34a093163cd527a0870fa75627b888715d8630a8a443ee6a6dff34d"} Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.760695 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerStarted","Data":"b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca"} Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.762583 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerStarted","Data":"a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461"} Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.767976 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kxfsf" podStartSLOduration=2.387132962 podStartE2EDuration="1m36.767965965s" podCreationTimestamp="2025-10-02 12:50:53 +0000 UTC" firstStartedPulling="2025-10-02 12:50:54.842356186 +0000 UTC m=+98.948767079" lastFinishedPulling="2025-10-02 12:52:29.223189189 +0000 UTC m=+193.329600082" observedRunningTime="2025-10-02 12:52:29.766230374 +0000 UTC m=+193.872641257" watchObservedRunningTime="2025-10-02 12:52:29.767965965 +0000 UTC m=+193.874376848" Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.790841 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tzk59" podStartSLOduration=3.489981454 podStartE2EDuration="1m37.790821969s" podCreationTimestamp="2025-10-02 12:50:52 +0000 UTC" firstStartedPulling="2025-10-02 12:50:54.86123262 +0000 UTC m=+98.967643503" lastFinishedPulling="2025-10-02 12:52:29.162073085 +0000 UTC m=+193.268484018" observedRunningTime="2025-10-02 12:52:29.78645082 +0000 UTC m=+193.892861723" watchObservedRunningTime="2025-10-02 12:52:29.790821969 +0000 UTC m=+193.897232852" Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.811372 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rzt2p" podStartSLOduration=4.869035899 podStartE2EDuration="1m34.811350195s" podCreationTimestamp="2025-10-02 12:50:55 +0000 UTC" firstStartedPulling="2025-10-02 12:50:57.97233569 +0000 UTC m=+102.078746573" lastFinishedPulling="2025-10-02 12:52:27.914649986 +0000 UTC m=+192.021060869" observedRunningTime="2025-10-02 12:52:29.809294145 +0000 UTC m=+193.915705028" watchObservedRunningTime="2025-10-02 12:52:29.811350195 +0000 UTC m=+193.917761088" Oct 02 12:52:29 crc kubenswrapper[4710]: I1002 12:52:29.927861 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.041237 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhsbt\" (UniqueName: \"kubernetes.io/projected/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-kube-api-access-vhsbt\") pod \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.041346 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-utilities\") pod \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.041494 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-catalog-content\") pod \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\" (UID: \"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d\") " Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.042076 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-utilities" (OuterVolumeSpecName: "utilities") pod "af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" (UID: "af1a2d7d-ccb8-4532-b6b1-2da89b14a32d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.048387 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-kube-api-access-vhsbt" (OuterVolumeSpecName: "kube-api-access-vhsbt") pod "af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" (UID: "af1a2d7d-ccb8-4532-b6b1-2da89b14a32d"). InnerVolumeSpecName "kube-api-access-vhsbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.142660 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhsbt\" (UniqueName: \"kubernetes.io/projected/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-kube-api-access-vhsbt\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.143106 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.145452 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" (UID: "af1a2d7d-ccb8-4532-b6b1-2da89b14a32d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.244795 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.771826 4710 generic.go:334] "Generic (PLEG): container finished" podID="8527e0f4-d514-45e1-89db-2d07c359387d" containerID="a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461" exitCode=0 Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.771863 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerDied","Data":"a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461"} Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.777456 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rsbdg" event={"ID":"af1a2d7d-ccb8-4532-b6b1-2da89b14a32d","Type":"ContainerDied","Data":"e6171239d42a0497fd701da9f814676f5562f5390c8ab29f49848ab4b155f466"} Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.777518 4710 scope.go:117] "RemoveContainer" containerID="861126c2c34a093163cd527a0870fa75627b888715d8630a8a443ee6a6dff34d" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.777684 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rsbdg" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.800934 4710 scope.go:117] "RemoveContainer" containerID="456cb6b188d03ff78d30a7d49370fc36603aec1ac8b7f28ece95bcb60e84d40e" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.814971 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rsbdg"] Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.823838 4710 scope.go:117] "RemoveContainer" containerID="2312b8d9a45c3a376732fe75660fa57b8acf33be62463e0baa0e7c78dde6315c" Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.825353 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rsbdg"] Oct 02 12:52:30 crc kubenswrapper[4710]: I1002 12:52:30.910911 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" path="/var/lib/kubelet/pods/af1a2d7d-ccb8-4532-b6b1-2da89b14a32d/volumes" Oct 02 12:52:31 crc kubenswrapper[4710]: I1002 12:52:31.786388 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerStarted","Data":"cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c"} Oct 02 12:52:31 crc kubenswrapper[4710]: I1002 12:52:31.801870 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kkj24" podStartSLOduration=3.298937902 podStartE2EDuration="1m39.801854986s" podCreationTimestamp="2025-10-02 12:50:52 +0000 UTC" firstStartedPulling="2025-10-02 12:50:54.85166385 +0000 UTC m=+98.958074743" lastFinishedPulling="2025-10-02 12:52:31.354580944 +0000 UTC m=+195.460991827" observedRunningTime="2025-10-02 12:52:31.799539569 +0000 UTC m=+195.905950452" watchObservedRunningTime="2025-10-02 12:52:31.801854986 +0000 UTC m=+195.908265859" Oct 02 12:52:32 crc kubenswrapper[4710]: I1002 12:52:32.784407 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:52:32 crc kubenswrapper[4710]: I1002 12:52:32.784700 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.259783 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.260141 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.325590 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.372724 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.372810 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.435995 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:52:33 crc kubenswrapper[4710]: I1002 12:52:33.836821 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kkj24" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="registry-server" probeResult="failure" output=< Oct 02 12:52:33 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 12:52:33 crc kubenswrapper[4710]: > Oct 02 12:52:35 crc kubenswrapper[4710]: I1002 12:52:35.012289 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 12:52:36 crc kubenswrapper[4710]: I1002 12:52:36.116833 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:52:36 crc kubenswrapper[4710]: I1002 12:52:36.117094 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:52:36 crc kubenswrapper[4710]: I1002 12:52:36.157997 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:52:36 crc kubenswrapper[4710]: I1002 12:52:36.860734 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:52:37 crc kubenswrapper[4710]: I1002 12:52:37.654713 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-284s4"] Oct 02 12:52:42 crc kubenswrapper[4710]: I1002 12:52:42.829125 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:52:42 crc kubenswrapper[4710]: I1002 12:52:42.869679 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:52:43 crc kubenswrapper[4710]: I1002 12:52:43.298970 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:52:43 crc kubenswrapper[4710]: I1002 12:52:43.411515 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:52:44 crc kubenswrapper[4710]: I1002 12:52:44.654708 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tzk59"] Oct 02 12:52:44 crc kubenswrapper[4710]: I1002 12:52:44.655204 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tzk59" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="registry-server" containerID="cri-o://d7535790593eca2803f42201a4983c8fd4d1bbed389f47e113911badfb9a9048" gracePeriod=2 Oct 02 12:52:44 crc kubenswrapper[4710]: I1002 12:52:44.872430 4710 generic.go:334] "Generic (PLEG): container finished" podID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerID="d7535790593eca2803f42201a4983c8fd4d1bbed389f47e113911badfb9a9048" exitCode=0 Oct 02 12:52:44 crc kubenswrapper[4710]: I1002 12:52:44.872474 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerDied","Data":"d7535790593eca2803f42201a4983c8fd4d1bbed389f47e113911badfb9a9048"} Oct 02 12:52:44 crc kubenswrapper[4710]: I1002 12:52:44.985842 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.130589 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-catalog-content\") pod \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.130644 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwr5z\" (UniqueName: \"kubernetes.io/projected/2ef28114-3b05-44a8-b1e5-5bbbda54d657-kube-api-access-hwr5z\") pod \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.130721 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-utilities\") pod \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\" (UID: \"2ef28114-3b05-44a8-b1e5-5bbbda54d657\") " Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.131720 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-utilities" (OuterVolumeSpecName: "utilities") pod "2ef28114-3b05-44a8-b1e5-5bbbda54d657" (UID: "2ef28114-3b05-44a8-b1e5-5bbbda54d657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.138975 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef28114-3b05-44a8-b1e5-5bbbda54d657-kube-api-access-hwr5z" (OuterVolumeSpecName: "kube-api-access-hwr5z") pod "2ef28114-3b05-44a8-b1e5-5bbbda54d657" (UID: "2ef28114-3b05-44a8-b1e5-5bbbda54d657"). InnerVolumeSpecName "kube-api-access-hwr5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.171517 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ef28114-3b05-44a8-b1e5-5bbbda54d657" (UID: "2ef28114-3b05-44a8-b1e5-5bbbda54d657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.231675 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.231729 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwr5z\" (UniqueName: \"kubernetes.io/projected/2ef28114-3b05-44a8-b1e5-5bbbda54d657-kube-api-access-hwr5z\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.231784 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef28114-3b05-44a8-b1e5-5bbbda54d657-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.656811 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxfsf"] Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.657047 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kxfsf" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="registry-server" containerID="cri-o://65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491" gracePeriod=2 Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.879209 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tzk59" event={"ID":"2ef28114-3b05-44a8-b1e5-5bbbda54d657","Type":"ContainerDied","Data":"9849009f33e28e82aed7b22c035301860563d53bbde2ab038ccc58ab1bf958cc"} Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.879267 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tzk59" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.879270 4710 scope.go:117] "RemoveContainer" containerID="d7535790593eca2803f42201a4983c8fd4d1bbed389f47e113911badfb9a9048" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.896211 4710 scope.go:117] "RemoveContainer" containerID="f2bf7535ab755fe5a77dd4d4766e605fc56405b5017885fe5b3096ae4f40d5d3" Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.907045 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tzk59"] Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.908642 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tzk59"] Oct 02 12:52:45 crc kubenswrapper[4710]: I1002 12:52:45.916400 4710 scope.go:117] "RemoveContainer" containerID="3c545cddabf9b8e898b9f6242e87b51a66564761fbaf3f7eb3f5d351a6f0fdf1" Oct 02 12:52:46 crc kubenswrapper[4710]: I1002 12:52:46.911178 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" path="/var/lib/kubelet/pods/2ef28114-3b05-44a8-b1e5-5bbbda54d657/volumes" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.491310 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.662602 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hpw6\" (UniqueName: \"kubernetes.io/projected/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-kube-api-access-6hpw6\") pod \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.662662 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-catalog-content\") pod \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.662829 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-utilities\") pod \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\" (UID: \"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5\") " Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.663678 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-utilities" (OuterVolumeSpecName: "utilities") pod "1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" (UID: "1535dfb1-9951-4d3e-ace9-2c70a4d0abb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.672816 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-kube-api-access-6hpw6" (OuterVolumeSpecName: "kube-api-access-6hpw6") pod "1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" (UID: "1535dfb1-9951-4d3e-ace9-2c70a4d0abb5"). InnerVolumeSpecName "kube-api-access-6hpw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.710617 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" (UID: "1535dfb1-9951-4d3e-ace9-2c70a4d0abb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.763696 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hpw6\" (UniqueName: \"kubernetes.io/projected/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-kube-api-access-6hpw6\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.763732 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.763741 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.890777 4710 generic.go:334] "Generic (PLEG): container finished" podID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerID="65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491" exitCode=0 Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.890821 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxfsf" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.890814 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxfsf" event={"ID":"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5","Type":"ContainerDied","Data":"65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491"} Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.891231 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxfsf" event={"ID":"1535dfb1-9951-4d3e-ace9-2c70a4d0abb5","Type":"ContainerDied","Data":"999517f3675887da960770e3751c143d6b3d884945f2398d8f0247ac5017b5f9"} Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.891319 4710 scope.go:117] "RemoveContainer" containerID="65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.914205 4710 scope.go:117] "RemoveContainer" containerID="9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.914346 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxfsf"] Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.917382 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kxfsf"] Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.944902 4710 scope.go:117] "RemoveContainer" containerID="60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.959440 4710 scope.go:117] "RemoveContainer" containerID="65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491" Oct 02 12:52:47 crc kubenswrapper[4710]: E1002 12:52:47.959738 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491\": container with ID starting with 65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491 not found: ID does not exist" containerID="65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.959790 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491"} err="failed to get container status \"65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491\": rpc error: code = NotFound desc = could not find container \"65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491\": container with ID starting with 65d6fc90a087885c29c16640e91a95def1e84b8bd082cdbf99f8015334782491 not found: ID does not exist" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.959817 4710 scope.go:117] "RemoveContainer" containerID="9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02" Oct 02 12:52:47 crc kubenswrapper[4710]: E1002 12:52:47.960077 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02\": container with ID starting with 9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02 not found: ID does not exist" containerID="9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.960099 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02"} err="failed to get container status \"9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02\": rpc error: code = NotFound desc = could not find container \"9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02\": container with ID starting with 9d1304298c12f3dcd5b6aa7e23ea47ba294ad15f1b04400be600a7162ac80a02 not found: ID does not exist" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.960114 4710 scope.go:117] "RemoveContainer" containerID="60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3" Oct 02 12:52:47 crc kubenswrapper[4710]: E1002 12:52:47.960284 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3\": container with ID starting with 60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3 not found: ID does not exist" containerID="60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3" Oct 02 12:52:47 crc kubenswrapper[4710]: I1002 12:52:47.960305 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3"} err="failed to get container status \"60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3\": rpc error: code = NotFound desc = could not find container \"60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3\": container with ID starting with 60f3bd815984fb4c0a38148ab2a8260ee344a61a3c2a63924d1efbf314997fc3 not found: ID does not exist" Oct 02 12:52:48 crc kubenswrapper[4710]: I1002 12:52:48.910313 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" path="/var/lib/kubelet/pods/1535dfb1-9951-4d3e-ace9-2c70a4d0abb5/volumes" Oct 02 12:52:52 crc kubenswrapper[4710]: I1002 12:52:52.530653 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:52:52 crc kubenswrapper[4710]: I1002 12:52:52.531053 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:52:52 crc kubenswrapper[4710]: I1002 12:52:52.531103 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:52:52 crc kubenswrapper[4710]: I1002 12:52:52.531645 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:52:52 crc kubenswrapper[4710]: I1002 12:52:52.531702 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c" gracePeriod=600 Oct 02 12:52:53 crc kubenswrapper[4710]: I1002 12:52:53.923566 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c" exitCode=0 Oct 02 12:52:53 crc kubenswrapper[4710]: I1002 12:52:53.923686 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c"} Oct 02 12:52:54 crc kubenswrapper[4710]: I1002 12:52:54.930929 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"d1f26514b5ee22a80a5d66c96251f5d904520e9599cd1deeb21c078bb6cf25a5"} Oct 02 12:53:02 crc kubenswrapper[4710]: I1002 12:53:02.683912 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" podUID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" containerName="oauth-openshift" containerID="cri-o://ba99bdaa823d37dc1177a2d212530b87cd79407ebdd695aea591733b71c2b4dc" gracePeriod=15 Oct 02 12:53:02 crc kubenswrapper[4710]: I1002 12:53:02.977473 4710 generic.go:334] "Generic (PLEG): container finished" podID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" containerID="ba99bdaa823d37dc1177a2d212530b87cd79407ebdd695aea591733b71c2b4dc" exitCode=0 Oct 02 12:53:02 crc kubenswrapper[4710]: I1002 12:53:02.977533 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" event={"ID":"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d","Type":"ContainerDied","Data":"ba99bdaa823d37dc1177a2d212530b87cd79407ebdd695aea591733b71c2b4dc"} Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.108592 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.139965 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-775b8f44b6-mjw4b"] Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140170 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8612401-3e78-458a-ad2d-746b2149d7c4" containerName="pruner" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140183 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8612401-3e78-458a-ad2d-746b2149d7c4" containerName="pruner" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140195 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140202 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140212 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140220 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140230 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140238 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140253 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140259 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140266 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140272 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140280 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140286 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140293 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140298 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140305 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140311 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140317 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140323 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140330 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140337 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="extract-utilities" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140343 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140349 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="extract-content" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140354 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" containerName="oauth-openshift" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140360 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" containerName="oauth-openshift" Oct 02 12:53:03 crc kubenswrapper[4710]: E1002 12:53:03.140367 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140373 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140463 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1a2d7d-ccb8-4532-b6b1-2da89b14a32d" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140477 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef28114-3b05-44a8-b1e5-5bbbda54d657" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140487 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="1535dfb1-9951-4d3e-ace9-2c70a4d0abb5" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140496 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8612401-3e78-458a-ad2d-746b2149d7c4" containerName="pruner" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140503 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" containerName="oauth-openshift" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140509 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab22687-b3c6-4efa-8926-bde00f550f82" containerName="registry-server" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.140902 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.147821 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-775b8f44b6-mjw4b"] Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274740 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-dir\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274811 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-error\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274836 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d64d2\" (UniqueName: \"kubernetes.io/projected/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-kube-api-access-d64d2\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274858 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-session\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274857 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274888 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-service-ca\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.274998 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-login\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275032 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-policies\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275061 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-router-certs\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275087 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-cliconfig\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275141 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-trusted-ca-bundle\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275180 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-idp-0-file-data\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275213 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-ocp-branding-template\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275237 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-serving-cert\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275273 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-provider-selection\") pod \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\" (UID: \"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d\") " Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275473 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-audit-dir\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275503 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-router-certs\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275593 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275623 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-login\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275691 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275725 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-audit-policies\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275794 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275843 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275871 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275893 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x726h\" (UniqueName: \"kubernetes.io/projected/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-kube-api-access-x726h\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275931 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-error\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275959 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-session\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.276012 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.276041 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-service-ca\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.276096 4710 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.275591 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.278307 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.278877 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.279046 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.281132 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.284389 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-kube-api-access-d64d2" (OuterVolumeSpecName: "kube-api-access-d64d2") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "kube-api-access-d64d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.284414 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.284659 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.284851 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.285413 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.285650 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.286313 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.289090 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" (UID: "ac9d2381-0ef0-46a4-bc0c-281fa429ef2d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377075 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-session\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377178 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377243 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-service-ca\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377309 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-audit-dir\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377354 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-router-certs\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377421 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377469 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-login\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377563 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377617 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-audit-policies\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377685 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377789 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-audit-dir\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377800 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377868 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377888 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x726h\" (UniqueName: \"kubernetes.io/projected/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-kube-api-access-x726h\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.377932 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-error\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378007 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378018 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378028 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378039 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378049 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d64d2\" (UniqueName: \"kubernetes.io/projected/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-kube-api-access-d64d2\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378059 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378159 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378381 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-service-ca\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378627 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378932 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.378990 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.379445 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-audit-policies\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.379627 4710 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.379666 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.379688 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.379708 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.379789 4710 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.381152 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-session\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.381249 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.381326 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-login\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.382792 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-router-certs\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.384281 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.384446 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.384508 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.385816 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-v4-0-config-user-template-error\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.393597 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x726h\" (UniqueName: \"kubernetes.io/projected/7934bd50-1cfb-4c49-95f7-61bb2e617ac3-kube-api-access-x726h\") pod \"oauth-openshift-775b8f44b6-mjw4b\" (UID: \"7934bd50-1cfb-4c49-95f7-61bb2e617ac3\") " pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.459096 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.915337 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-775b8f44b6-mjw4b"] Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.989326 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" event={"ID":"7934bd50-1cfb-4c49-95f7-61bb2e617ac3","Type":"ContainerStarted","Data":"129527c88a8d97a4a20323edfbda8cfa314314686664e037a2583778b0c5eb1f"} Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.990495 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" event={"ID":"ac9d2381-0ef0-46a4-bc0c-281fa429ef2d","Type":"ContainerDied","Data":"36eee9a1a5e9a7990a5ed8c085df27fc30ea349f0f50f229278427e08b23299e"} Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.990534 4710 scope.go:117] "RemoveContainer" containerID="ba99bdaa823d37dc1177a2d212530b87cd79407ebdd695aea591733b71c2b4dc" Oct 02 12:53:03 crc kubenswrapper[4710]: I1002 12:53:03.990634 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-284s4" Oct 02 12:53:04 crc kubenswrapper[4710]: I1002 12:53:04.018580 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-284s4"] Oct 02 12:53:04 crc kubenswrapper[4710]: I1002 12:53:04.021061 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-284s4"] Oct 02 12:53:04 crc kubenswrapper[4710]: I1002 12:53:04.909991 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac9d2381-0ef0-46a4-bc0c-281fa429ef2d" path="/var/lib/kubelet/pods/ac9d2381-0ef0-46a4-bc0c-281fa429ef2d/volumes" Oct 02 12:53:04 crc kubenswrapper[4710]: I1002 12:53:04.998509 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" event={"ID":"7934bd50-1cfb-4c49-95f7-61bb2e617ac3","Type":"ContainerStarted","Data":"497676c95bb2a55fb3792e5fb5dffe9fc332b0c40adb191c3487dc94c4bcdffa"} Oct 02 12:53:05 crc kubenswrapper[4710]: I1002 12:53:04.999378 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:05 crc kubenswrapper[4710]: I1002 12:53:05.004898 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" Oct 02 12:53:05 crc kubenswrapper[4710]: I1002 12:53:05.046691 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-775b8f44b6-mjw4b" podStartSLOduration=28.04667249 podStartE2EDuration="28.04667249s" podCreationTimestamp="2025-10-02 12:52:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:53:05.018001896 +0000 UTC m=+229.124412799" watchObservedRunningTime="2025-10-02 12:53:05.04667249 +0000 UTC m=+229.153083383" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.343121 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kkj24"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.343925 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kkj24" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="registry-server" containerID="cri-o://cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c" gracePeriod=30 Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.354551 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7tgfs"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.354858 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7tgfs" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="registry-server" containerID="cri-o://4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc" gracePeriod=30 Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.374553 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cfvqq"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.375061 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" containerID="cri-o://c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453" gracePeriod=30 Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.387814 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vwzwf"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.388155 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vwzwf" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="registry-server" containerID="cri-o://ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2" gracePeriod=30 Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.391005 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z2qt5"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.391980 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.406666 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rzt2p"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.406931 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rzt2p" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="registry-server" containerID="cri-o://b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca" gracePeriod=30 Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.416115 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z2qt5"] Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.452017 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqct9\" (UniqueName: \"kubernetes.io/projected/a54e807e-bc57-41a3-8298-da6597e3e669-kube-api-access-sqct9\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.452081 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a54e807e-bc57-41a3-8298-da6597e3e669-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.452252 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a54e807e-bc57-41a3-8298-da6597e3e669-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.553810 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqct9\" (UniqueName: \"kubernetes.io/projected/a54e807e-bc57-41a3-8298-da6597e3e669-kube-api-access-sqct9\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.554264 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a54e807e-bc57-41a3-8298-da6597e3e669-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.554407 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a54e807e-bc57-41a3-8298-da6597e3e669-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.556896 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a54e807e-bc57-41a3-8298-da6597e3e669-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.561987 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a54e807e-bc57-41a3-8298-da6597e3e669-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.573082 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqct9\" (UniqueName: \"kubernetes.io/projected/a54e807e-bc57-41a3-8298-da6597e3e669-kube-api-access-sqct9\") pod \"marketplace-operator-79b997595-z2qt5\" (UID: \"a54e807e-bc57-41a3-8298-da6597e3e669\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.710047 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.827428 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.831384 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.862849 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.883422 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.894238 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.959844 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-operator-metrics\") pod \"c125db0e-2362-4203-88c9-26a1f24c926c\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.960108 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-utilities\") pod \"8527e0f4-d514-45e1-89db-2d07c359387d\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.961844 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-utilities" (OuterVolumeSpecName: "utilities") pod "8527e0f4-d514-45e1-89db-2d07c359387d" (UID: "8527e0f4-d514-45e1-89db-2d07c359387d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.961875 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-utilities\") pod \"0ec33244-0383-45c4-aecc-17d4c3f4435c\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962075 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pvsm\" (UniqueName: \"kubernetes.io/projected/8527e0f4-d514-45e1-89db-2d07c359387d-kube-api-access-8pvsm\") pod \"8527e0f4-d514-45e1-89db-2d07c359387d\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962099 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-utilities\") pod \"567598d7-6c35-46b0-bd80-5c04b4f7e024\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962149 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qks7\" (UniqueName: \"kubernetes.io/projected/0ec33244-0383-45c4-aecc-17d4c3f4435c-kube-api-access-5qks7\") pod \"0ec33244-0383-45c4-aecc-17d4c3f4435c\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962178 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-trusted-ca\") pod \"c125db0e-2362-4203-88c9-26a1f24c926c\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962215 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-catalog-content\") pod \"07062fef-42ef-410d-adc5-0ce1063b6f26\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962234 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf9nz\" (UniqueName: \"kubernetes.io/projected/07062fef-42ef-410d-adc5-0ce1063b6f26-kube-api-access-sf9nz\") pod \"07062fef-42ef-410d-adc5-0ce1063b6f26\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962251 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-utilities\") pod \"07062fef-42ef-410d-adc5-0ce1063b6f26\" (UID: \"07062fef-42ef-410d-adc5-0ce1063b6f26\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962303 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-catalog-content\") pod \"8527e0f4-d514-45e1-89db-2d07c359387d\" (UID: \"8527e0f4-d514-45e1-89db-2d07c359387d\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962343 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-catalog-content\") pod \"0ec33244-0383-45c4-aecc-17d4c3f4435c\" (UID: \"0ec33244-0383-45c4-aecc-17d4c3f4435c\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962388 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-catalog-content\") pod \"567598d7-6c35-46b0-bd80-5c04b4f7e024\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962415 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmlxz\" (UniqueName: \"kubernetes.io/projected/c125db0e-2362-4203-88c9-26a1f24c926c-kube-api-access-vmlxz\") pod \"c125db0e-2362-4203-88c9-26a1f24c926c\" (UID: \"c125db0e-2362-4203-88c9-26a1f24c926c\") " Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.962555 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-utilities" (OuterVolumeSpecName: "utilities") pod "0ec33244-0383-45c4-aecc-17d4c3f4435c" (UID: "0ec33244-0383-45c4-aecc-17d4c3f4435c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.963231 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c125db0e-2362-4203-88c9-26a1f24c926c" (UID: "c125db0e-2362-4203-88c9-26a1f24c926c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.963306 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-utilities" (OuterVolumeSpecName: "utilities") pod "567598d7-6c35-46b0-bd80-5c04b4f7e024" (UID: "567598d7-6c35-46b0-bd80-5c04b4f7e024"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.963512 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.963531 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.963544 4710 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.963555 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.964113 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-utilities" (OuterVolumeSpecName: "utilities") pod "07062fef-42ef-410d-adc5-0ce1063b6f26" (UID: "07062fef-42ef-410d-adc5-0ce1063b6f26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.965851 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8527e0f4-d514-45e1-89db-2d07c359387d-kube-api-access-8pvsm" (OuterVolumeSpecName: "kube-api-access-8pvsm") pod "8527e0f4-d514-45e1-89db-2d07c359387d" (UID: "8527e0f4-d514-45e1-89db-2d07c359387d"). InnerVolumeSpecName "kube-api-access-8pvsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.967427 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c125db0e-2362-4203-88c9-26a1f24c926c-kube-api-access-vmlxz" (OuterVolumeSpecName: "kube-api-access-vmlxz") pod "c125db0e-2362-4203-88c9-26a1f24c926c" (UID: "c125db0e-2362-4203-88c9-26a1f24c926c"). InnerVolumeSpecName "kube-api-access-vmlxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.968994 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c125db0e-2362-4203-88c9-26a1f24c926c" (UID: "c125db0e-2362-4203-88c9-26a1f24c926c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.973195 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07062fef-42ef-410d-adc5-0ce1063b6f26-kube-api-access-sf9nz" (OuterVolumeSpecName: "kube-api-access-sf9nz") pod "07062fef-42ef-410d-adc5-0ce1063b6f26" (UID: "07062fef-42ef-410d-adc5-0ce1063b6f26"). InnerVolumeSpecName "kube-api-access-sf9nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:53:23 crc kubenswrapper[4710]: I1002 12:53:23.984222 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z2qt5"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.002901 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec33244-0383-45c4-aecc-17d4c3f4435c-kube-api-access-5qks7" (OuterVolumeSpecName: "kube-api-access-5qks7") pod "0ec33244-0383-45c4-aecc-17d4c3f4435c" (UID: "0ec33244-0383-45c4-aecc-17d4c3f4435c"). InnerVolumeSpecName "kube-api-access-5qks7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.003509 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ec33244-0383-45c4-aecc-17d4c3f4435c" (UID: "0ec33244-0383-45c4-aecc-17d4c3f4435c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.044517 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "567598d7-6c35-46b0-bd80-5c04b4f7e024" (UID: "567598d7-6c35-46b0-bd80-5c04b4f7e024"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.051952 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07062fef-42ef-410d-adc5-0ce1063b6f26" (UID: "07062fef-42ef-410d-adc5-0ce1063b6f26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065042 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhsw8\" (UniqueName: \"kubernetes.io/projected/567598d7-6c35-46b0-bd80-5c04b4f7e024-kube-api-access-bhsw8\") pod \"567598d7-6c35-46b0-bd80-5c04b4f7e024\" (UID: \"567598d7-6c35-46b0-bd80-5c04b4f7e024\") " Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065404 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/567598d7-6c35-46b0-bd80-5c04b4f7e024-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065433 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmlxz\" (UniqueName: \"kubernetes.io/projected/c125db0e-2362-4203-88c9-26a1f24c926c-kube-api-access-vmlxz\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065454 4710 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c125db0e-2362-4203-88c9-26a1f24c926c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065474 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pvsm\" (UniqueName: \"kubernetes.io/projected/8527e0f4-d514-45e1-89db-2d07c359387d-kube-api-access-8pvsm\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065488 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qks7\" (UniqueName: \"kubernetes.io/projected/0ec33244-0383-45c4-aecc-17d4c3f4435c-kube-api-access-5qks7\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065497 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065508 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf9nz\" (UniqueName: \"kubernetes.io/projected/07062fef-42ef-410d-adc5-0ce1063b6f26-kube-api-access-sf9nz\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065516 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07062fef-42ef-410d-adc5-0ce1063b6f26-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.065526 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec33244-0383-45c4-aecc-17d4c3f4435c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.067734 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567598d7-6c35-46b0-bd80-5c04b4f7e024-kube-api-access-bhsw8" (OuterVolumeSpecName: "kube-api-access-bhsw8") pod "567598d7-6c35-46b0-bd80-5c04b4f7e024" (UID: "567598d7-6c35-46b0-bd80-5c04b4f7e024"). InnerVolumeSpecName "kube-api-access-bhsw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.070936 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8527e0f4-d514-45e1-89db-2d07c359387d" (UID: "8527e0f4-d514-45e1-89db-2d07c359387d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.108474 4710 generic.go:334] "Generic (PLEG): container finished" podID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerID="4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc" exitCode=0 Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.108537 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tgfs" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.108650 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerDied","Data":"4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.108707 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tgfs" event={"ID":"567598d7-6c35-46b0-bd80-5c04b4f7e024","Type":"ContainerDied","Data":"cf4366eab9a08e12a5593e952fa552a662878dea81a174ba3e0b2bb58875bf05"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.108730 4710 scope.go:117] "RemoveContainer" containerID="4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.110997 4710 generic.go:334] "Generic (PLEG): container finished" podID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerID="ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2" exitCode=0 Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.111151 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vwzwf" event={"ID":"0ec33244-0383-45c4-aecc-17d4c3f4435c","Type":"ContainerDied","Data":"ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.111260 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vwzwf" event={"ID":"0ec33244-0383-45c4-aecc-17d4c3f4435c","Type":"ContainerDied","Data":"a4a3425314af0b543b1fa2bc803bc70810960c74a1fc25bea6aeda693c8542d6"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.111066 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vwzwf" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.113140 4710 generic.go:334] "Generic (PLEG): container finished" podID="c125db0e-2362-4203-88c9-26a1f24c926c" containerID="c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453" exitCode=0 Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.113238 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.113413 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" event={"ID":"c125db0e-2362-4203-88c9-26a1f24c926c","Type":"ContainerDied","Data":"c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.113500 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cfvqq" event={"ID":"c125db0e-2362-4203-88c9-26a1f24c926c","Type":"ContainerDied","Data":"686bab809f0704f3ecea6a0b7d17bf203a8bac1795fc840c90e45322cd06435a"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.115791 4710 generic.go:334] "Generic (PLEG): container finished" podID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerID="b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca" exitCode=0 Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.115836 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerDied","Data":"b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.115855 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzt2p" event={"ID":"07062fef-42ef-410d-adc5-0ce1063b6f26","Type":"ContainerDied","Data":"31730b266111f49d9b701fb3253f0e92134b38449eb896f8c9a18330d900dc78"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.115901 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzt2p" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.124681 4710 scope.go:117] "RemoveContainer" containerID="6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.129800 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" event={"ID":"a54e807e-bc57-41a3-8298-da6597e3e669","Type":"ContainerStarted","Data":"43ca54c04c7ab600e182a4409b8b69f91989383fa91e86e629144f419a143d37"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.133789 4710 generic.go:334] "Generic (PLEG): container finished" podID="8527e0f4-d514-45e1-89db-2d07c359387d" containerID="cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c" exitCode=0 Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.133856 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerDied","Data":"cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.133859 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkj24" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.133885 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkj24" event={"ID":"8527e0f4-d514-45e1-89db-2d07c359387d","Type":"ContainerDied","Data":"d2d13b545b44e820836dfa0d6e88ac777a7259a59cbb66ff5a313b9e2a3c3fb4"} Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.146578 4710 scope.go:117] "RemoveContainer" containerID="0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.149276 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rzt2p"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.156586 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rzt2p"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.163857 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7tgfs"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.167070 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8527e0f4-d514-45e1-89db-2d07c359387d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.167099 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhsw8\" (UniqueName: \"kubernetes.io/projected/567598d7-6c35-46b0-bd80-5c04b4f7e024-kube-api-access-bhsw8\") on node \"crc\" DevicePath \"\"" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.173365 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7tgfs"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.177623 4710 scope.go:117] "RemoveContainer" containerID="4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.180270 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc\": container with ID starting with 4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc not found: ID does not exist" containerID="4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.180481 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc"} err="failed to get container status \"4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc\": rpc error: code = NotFound desc = could not find container \"4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc\": container with ID starting with 4e9610426bd3363b14028dfbfbe4e9c357defefee57d2c55dcab0d256cdf3bbc not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.180622 4710 scope.go:117] "RemoveContainer" containerID="6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.180873 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vwzwf"] Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.181893 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667\": container with ID starting with 6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667 not found: ID does not exist" containerID="6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.182224 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667"} err="failed to get container status \"6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667\": rpc error: code = NotFound desc = could not find container \"6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667\": container with ID starting with 6829240aceb7e25387750a84c2af243409ea188fc135f6102e8292520707c667 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.182367 4710 scope.go:117] "RemoveContainer" containerID="0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.182908 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2\": container with ID starting with 0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2 not found: ID does not exist" containerID="0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.183080 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2"} err="failed to get container status \"0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2\": rpc error: code = NotFound desc = could not find container \"0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2\": container with ID starting with 0e01c700862d98eab43541e81a6e0bbd6d064a6d2936738ff7993102474396c2 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.184865 4710 scope.go:117] "RemoveContainer" containerID="ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.189709 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vwzwf"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.193536 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cfvqq"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.196248 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cfvqq"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.204378 4710 scope.go:117] "RemoveContainer" containerID="0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.204708 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kkj24"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.213985 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kkj24"] Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.225492 4710 scope.go:117] "RemoveContainer" containerID="6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.237646 4710 scope.go:117] "RemoveContainer" containerID="ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.238126 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2\": container with ID starting with ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2 not found: ID does not exist" containerID="ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.238159 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2"} err="failed to get container status \"ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2\": rpc error: code = NotFound desc = could not find container \"ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2\": container with ID starting with ab7bde95d0b82630fe08a2c9e024786ed2a0d9621da20c4da95862b039dc8fa2 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.238181 4710 scope.go:117] "RemoveContainer" containerID="0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.238392 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61\": container with ID starting with 0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61 not found: ID does not exist" containerID="0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.238412 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61"} err="failed to get container status \"0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61\": rpc error: code = NotFound desc = could not find container \"0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61\": container with ID starting with 0c131f426d7144d16eef955848107d440d2d2bf90c79a71468fae1e73d82cf61 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.238423 4710 scope.go:117] "RemoveContainer" containerID="6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.238666 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7\": container with ID starting with 6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7 not found: ID does not exist" containerID="6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.238685 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7"} err="failed to get container status \"6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7\": rpc error: code = NotFound desc = could not find container \"6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7\": container with ID starting with 6b2f2ba141b29b06da16d10305a1e57a82b90a227238fc67f07550e800a846e7 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.238696 4710 scope.go:117] "RemoveContainer" containerID="c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.249461 4710 scope.go:117] "RemoveContainer" containerID="c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.249706 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453\": container with ID starting with c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453 not found: ID does not exist" containerID="c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.249734 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453"} err="failed to get container status \"c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453\": rpc error: code = NotFound desc = could not find container \"c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453\": container with ID starting with c6459640a130d936d174d0009da8a4095af191b2d0fe5e7e706c11dd1becf453 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.249818 4710 scope.go:117] "RemoveContainer" containerID="b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.261019 4710 scope.go:117] "RemoveContainer" containerID="d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.271518 4710 scope.go:117] "RemoveContainer" containerID="fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.282361 4710 scope.go:117] "RemoveContainer" containerID="b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.282741 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca\": container with ID starting with b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca not found: ID does not exist" containerID="b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.282800 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca"} err="failed to get container status \"b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca\": rpc error: code = NotFound desc = could not find container \"b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca\": container with ID starting with b5b30f21553b1fcc110e02ad8e4abd64f19c6f3a5815d86bc5d47a2ff1677bca not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.282829 4710 scope.go:117] "RemoveContainer" containerID="d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.283053 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d\": container with ID starting with d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d not found: ID does not exist" containerID="d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.283082 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d"} err="failed to get container status \"d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d\": rpc error: code = NotFound desc = could not find container \"d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d\": container with ID starting with d6a8047e06078cf646d10ba7172f738ed165fb3387b12e0ad3890b468ec1328d not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.283100 4710 scope.go:117] "RemoveContainer" containerID="fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.283306 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606\": container with ID starting with fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606 not found: ID does not exist" containerID="fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.283333 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606"} err="failed to get container status \"fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606\": rpc error: code = NotFound desc = could not find container \"fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606\": container with ID starting with fcdbabca01a7687008ed1487ffdb6d793a8a1d975cb70ac5f2b1e7d3d7e19606 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.283348 4710 scope.go:117] "RemoveContainer" containerID="cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.300833 4710 scope.go:117] "RemoveContainer" containerID="a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.313709 4710 scope.go:117] "RemoveContainer" containerID="30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.326262 4710 scope.go:117] "RemoveContainer" containerID="cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.326822 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c\": container with ID starting with cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c not found: ID does not exist" containerID="cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.326861 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c"} err="failed to get container status \"cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c\": rpc error: code = NotFound desc = could not find container \"cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c\": container with ID starting with cb54eb4e22f063e514d6496de14c16d2dfd3d6b48d6d2ccd916550758dc92e4c not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.326888 4710 scope.go:117] "RemoveContainer" containerID="a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.327595 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461\": container with ID starting with a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461 not found: ID does not exist" containerID="a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.327618 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461"} err="failed to get container status \"a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461\": rpc error: code = NotFound desc = could not find container \"a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461\": container with ID starting with a4834e04419f0d938b3718a759bcf326b0220e8ba09ca93bcc3cc2a233a96461 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.327634 4710 scope.go:117] "RemoveContainer" containerID="30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4" Oct 02 12:53:24 crc kubenswrapper[4710]: E1002 12:53:24.327876 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4\": container with ID starting with 30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4 not found: ID does not exist" containerID="30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.327903 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4"} err="failed to get container status \"30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4\": rpc error: code = NotFound desc = could not find container \"30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4\": container with ID starting with 30ee31b08d7291990a50f74c6079c0a0802e94ff6f2fd910bf802c6140d174c4 not found: ID does not exist" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.913999 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" path="/var/lib/kubelet/pods/07062fef-42ef-410d-adc5-0ce1063b6f26/volumes" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.914727 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" path="/var/lib/kubelet/pods/0ec33244-0383-45c4-aecc-17d4c3f4435c/volumes" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.915476 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" path="/var/lib/kubelet/pods/567598d7-6c35-46b0-bd80-5c04b4f7e024/volumes" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.916695 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" path="/var/lib/kubelet/pods/8527e0f4-d514-45e1-89db-2d07c359387d/volumes" Oct 02 12:53:24 crc kubenswrapper[4710]: I1002 12:53:24.917419 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" path="/var/lib/kubelet/pods/c125db0e-2362-4203-88c9-26a1f24c926c/volumes" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.142084 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" event={"ID":"a54e807e-bc57-41a3-8298-da6597e3e669","Type":"ContainerStarted","Data":"8b1433f76f2443bf6217f48edb318c9ba7080b48bd7d196936f10e8c1ff351a9"} Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.142492 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.146051 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.161548 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z2qt5" podStartSLOduration=2.161518689 podStartE2EDuration="2.161518689s" podCreationTimestamp="2025-10-02 12:53:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:53:25.158937964 +0000 UTC m=+249.265348857" watchObservedRunningTime="2025-10-02 12:53:25.161518689 +0000 UTC m=+249.267929572" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.565912 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8scqr"] Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566153 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566168 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566177 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566185 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566198 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566205 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566213 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566220 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566230 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566237 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566249 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566256 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566267 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566276 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566284 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566291 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566301 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566308 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566320 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566327 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566339 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566347 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="extract-utilities" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566356 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566364 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="extract-content" Oct 02 12:53:25 crc kubenswrapper[4710]: E1002 12:53:25.566371 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566377 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566488 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="567598d7-6c35-46b0-bd80-5c04b4f7e024" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566504 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec33244-0383-45c4-aecc-17d4c3f4435c" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566517 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c125db0e-2362-4203-88c9-26a1f24c926c" containerName="marketplace-operator" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566530 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8527e0f4-d514-45e1-89db-2d07c359387d" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.566541 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="07062fef-42ef-410d-adc5-0ce1063b6f26" containerName="registry-server" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.567410 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.577217 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.581147 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8scqr"] Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.583329 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0832-a94c-47c9-853b-beee2115ce82-catalog-content\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.583395 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0832-a94c-47c9-853b-beee2115ce82-utilities\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.583484 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twtwn\" (UniqueName: \"kubernetes.io/projected/cc1a0832-a94c-47c9-853b-beee2115ce82-kube-api-access-twtwn\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.684133 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0832-a94c-47c9-853b-beee2115ce82-utilities\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.684220 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twtwn\" (UniqueName: \"kubernetes.io/projected/cc1a0832-a94c-47c9-853b-beee2115ce82-kube-api-access-twtwn\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.684262 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0832-a94c-47c9-853b-beee2115ce82-catalog-content\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.684629 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0832-a94c-47c9-853b-beee2115ce82-utilities\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.684812 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0832-a94c-47c9-853b-beee2115ce82-catalog-content\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.701050 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twtwn\" (UniqueName: \"kubernetes.io/projected/cc1a0832-a94c-47c9-853b-beee2115ce82-kube-api-access-twtwn\") pod \"certified-operators-8scqr\" (UID: \"cc1a0832-a94c-47c9-853b-beee2115ce82\") " pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.757926 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lb6tp"] Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.760543 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.764815 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.767858 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb6tp"] Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.899770 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.915614 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxlmf\" (UniqueName: \"kubernetes.io/projected/2e0e94c3-5165-4c97-a959-b5a26070fa99-kube-api-access-lxlmf\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.915678 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0e94c3-5165-4c97-a959-b5a26070fa99-utilities\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:25 crc kubenswrapper[4710]: I1002 12:53:25.915964 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0e94c3-5165-4c97-a959-b5a26070fa99-catalog-content\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.017003 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxlmf\" (UniqueName: \"kubernetes.io/projected/2e0e94c3-5165-4c97-a959-b5a26070fa99-kube-api-access-lxlmf\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.017390 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0e94c3-5165-4c97-a959-b5a26070fa99-utilities\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.017438 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0e94c3-5165-4c97-a959-b5a26070fa99-catalog-content\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.017934 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0e94c3-5165-4c97-a959-b5a26070fa99-utilities\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.018060 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0e94c3-5165-4c97-a959-b5a26070fa99-catalog-content\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.049702 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxlmf\" (UniqueName: \"kubernetes.io/projected/2e0e94c3-5165-4c97-a959-b5a26070fa99-kube-api-access-lxlmf\") pod \"redhat-marketplace-lb6tp\" (UID: \"2e0e94c3-5165-4c97-a959-b5a26070fa99\") " pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.094019 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8scqr"] Oct 02 12:53:26 crc kubenswrapper[4710]: W1002 12:53:26.098651 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc1a0832_a94c_47c9_853b_beee2115ce82.slice/crio-925da7a7ac239b28129f29866576cd43917e112032e82886fcf48b09c4ab1eaf WatchSource:0}: Error finding container 925da7a7ac239b28129f29866576cd43917e112032e82886fcf48b09c4ab1eaf: Status 404 returned error can't find the container with id 925da7a7ac239b28129f29866576cd43917e112032e82886fcf48b09c4ab1eaf Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.131249 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.163982 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8scqr" event={"ID":"cc1a0832-a94c-47c9-853b-beee2115ce82","Type":"ContainerStarted","Data":"925da7a7ac239b28129f29866576cd43917e112032e82886fcf48b09c4ab1eaf"} Oct 02 12:53:26 crc kubenswrapper[4710]: I1002 12:53:26.298614 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb6tp"] Oct 02 12:53:26 crc kubenswrapper[4710]: W1002 12:53:26.304178 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e0e94c3_5165_4c97_a959_b5a26070fa99.slice/crio-2dc9243e1c888ae67b866bff334c3c0b3bbeabda597b1b2dc128fe2126b31480 WatchSource:0}: Error finding container 2dc9243e1c888ae67b866bff334c3c0b3bbeabda597b1b2dc128fe2126b31480: Status 404 returned error can't find the container with id 2dc9243e1c888ae67b866bff334c3c0b3bbeabda597b1b2dc128fe2126b31480 Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.167882 4710 generic.go:334] "Generic (PLEG): container finished" podID="cc1a0832-a94c-47c9-853b-beee2115ce82" containerID="79c2d0a362b47a87fbbb5a3d85d14d19d52a6902e7f964ad44ee61033ec73891" exitCode=0 Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.167935 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8scqr" event={"ID":"cc1a0832-a94c-47c9-853b-beee2115ce82","Type":"ContainerDied","Data":"79c2d0a362b47a87fbbb5a3d85d14d19d52a6902e7f964ad44ee61033ec73891"} Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.170277 4710 generic.go:334] "Generic (PLEG): container finished" podID="2e0e94c3-5165-4c97-a959-b5a26070fa99" containerID="4f4233fc33217ce5491b363e8ba6c12f1ed52182528832f997717d1422c94ce0" exitCode=0 Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.170336 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb6tp" event={"ID":"2e0e94c3-5165-4c97-a959-b5a26070fa99","Type":"ContainerDied","Data":"4f4233fc33217ce5491b363e8ba6c12f1ed52182528832f997717d1422c94ce0"} Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.170395 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb6tp" event={"ID":"2e0e94c3-5165-4c97-a959-b5a26070fa99","Type":"ContainerStarted","Data":"2dc9243e1c888ae67b866bff334c3c0b3bbeabda597b1b2dc128fe2126b31480"} Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.962550 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hqkkb"] Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.964787 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.967439 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 12:53:27 crc kubenswrapper[4710]: I1002 12:53:27.979456 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqkkb"] Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.046012 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad1040-7c7d-4265-b428-d95694c6c346-utilities\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.046088 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad1040-7c7d-4265-b428-d95694c6c346-catalog-content\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.046404 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpzz\" (UniqueName: \"kubernetes.io/projected/87ad1040-7c7d-4265-b428-d95694c6c346-kube-api-access-xmpzz\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.148242 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpzz\" (UniqueName: \"kubernetes.io/projected/87ad1040-7c7d-4265-b428-d95694c6c346-kube-api-access-xmpzz\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.148369 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad1040-7c7d-4265-b428-d95694c6c346-utilities\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.148428 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad1040-7c7d-4265-b428-d95694c6c346-catalog-content\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.149194 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad1040-7c7d-4265-b428-d95694c6c346-utilities\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.149435 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad1040-7c7d-4265-b428-d95694c6c346-catalog-content\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.166551 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tm6bc"] Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.167602 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.169496 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.171808 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tm6bc"] Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.191926 4710 generic.go:334] "Generic (PLEG): container finished" podID="2e0e94c3-5165-4c97-a959-b5a26070fa99" containerID="be1b0a83da9a55c8fca0efe2c487c664d69c99ec8b5d31e286cc9067a9b8a5fa" exitCode=0 Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.191969 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb6tp" event={"ID":"2e0e94c3-5165-4c97-a959-b5a26070fa99","Type":"ContainerDied","Data":"be1b0a83da9a55c8fca0efe2c487c664d69c99ec8b5d31e286cc9067a9b8a5fa"} Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.193352 4710 generic.go:334] "Generic (PLEG): container finished" podID="cc1a0832-a94c-47c9-853b-beee2115ce82" containerID="ec63f4f04095dc0cddbbbd6700121bc75dd5c8e66eb5727396f61b611f91d220" exitCode=0 Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.193379 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8scqr" event={"ID":"cc1a0832-a94c-47c9-853b-beee2115ce82","Type":"ContainerDied","Data":"ec63f4f04095dc0cddbbbd6700121bc75dd5c8e66eb5727396f61b611f91d220"} Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.236785 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpzz\" (UniqueName: \"kubernetes.io/projected/87ad1040-7c7d-4265-b428-d95694c6c346-kube-api-access-xmpzz\") pod \"community-operators-hqkkb\" (UID: \"87ad1040-7c7d-4265-b428-d95694c6c346\") " pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.249498 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-utilities\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.249618 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87pcs\" (UniqueName: \"kubernetes.io/projected/fccaa30c-236a-4f2a-a2de-644399d3679d-kube-api-access-87pcs\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.249663 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-catalog-content\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.287192 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.350319 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87pcs\" (UniqueName: \"kubernetes.io/projected/fccaa30c-236a-4f2a-a2de-644399d3679d-kube-api-access-87pcs\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.350390 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-catalog-content\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.350424 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-utilities\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.350904 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-utilities\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.351015 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-catalog-content\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.369121 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87pcs\" (UniqueName: \"kubernetes.io/projected/fccaa30c-236a-4f2a-a2de-644399d3679d-kube-api-access-87pcs\") pod \"redhat-operators-tm6bc\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.581129 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.668857 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqkkb"] Oct 02 12:53:28 crc kubenswrapper[4710]: I1002 12:53:28.796847 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tm6bc"] Oct 02 12:53:28 crc kubenswrapper[4710]: W1002 12:53:28.817869 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfccaa30c_236a_4f2a_a2de_644399d3679d.slice/crio-85be7e47c784873282881eca6ea1e898b9d21b5de9928d391582be8c921d7d97 WatchSource:0}: Error finding container 85be7e47c784873282881eca6ea1e898b9d21b5de9928d391582be8c921d7d97: Status 404 returned error can't find the container with id 85be7e47c784873282881eca6ea1e898b9d21b5de9928d391582be8c921d7d97 Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.201509 4710 generic.go:334] "Generic (PLEG): container finished" podID="87ad1040-7c7d-4265-b428-d95694c6c346" containerID="d178e4768fc486b94a68da5b6a6c96ba5f8b10e0a584ee8d7a5c12af2aec770a" exitCode=0 Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.201585 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqkkb" event={"ID":"87ad1040-7c7d-4265-b428-d95694c6c346","Type":"ContainerDied","Data":"d178e4768fc486b94a68da5b6a6c96ba5f8b10e0a584ee8d7a5c12af2aec770a"} Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.201615 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqkkb" event={"ID":"87ad1040-7c7d-4265-b428-d95694c6c346","Type":"ContainerStarted","Data":"eca47fc1cfac045f17a433d29fdc70428e678ceda151dc40ecd219644d8f4807"} Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.204615 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8scqr" event={"ID":"cc1a0832-a94c-47c9-853b-beee2115ce82","Type":"ContainerStarted","Data":"5c6690be3a30d3ab23339e707320bda2d4bdce620eb76b435072f624954df740"} Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.210702 4710 generic.go:334] "Generic (PLEG): container finished" podID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerID="d36585f44fd80401717018a5c1fc798721d30594718dd2533a2159c68a343b7a" exitCode=0 Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.210773 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm6bc" event={"ID":"fccaa30c-236a-4f2a-a2de-644399d3679d","Type":"ContainerDied","Data":"d36585f44fd80401717018a5c1fc798721d30594718dd2533a2159c68a343b7a"} Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.210808 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm6bc" event={"ID":"fccaa30c-236a-4f2a-a2de-644399d3679d","Type":"ContainerStarted","Data":"85be7e47c784873282881eca6ea1e898b9d21b5de9928d391582be8c921d7d97"} Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.213885 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb6tp" event={"ID":"2e0e94c3-5165-4c97-a959-b5a26070fa99","Type":"ContainerStarted","Data":"86c1363376642ad5e0f8cc77da2ac019497a572bfa97968f8672047626a3ef45"} Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.240091 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lb6tp" podStartSLOduration=2.737951773 podStartE2EDuration="4.240072134s" podCreationTimestamp="2025-10-02 12:53:25 +0000 UTC" firstStartedPulling="2025-10-02 12:53:27.171489284 +0000 UTC m=+251.277900187" lastFinishedPulling="2025-10-02 12:53:28.673609675 +0000 UTC m=+252.780020548" observedRunningTime="2025-10-02 12:53:29.236994615 +0000 UTC m=+253.343405518" watchObservedRunningTime="2025-10-02 12:53:29.240072134 +0000 UTC m=+253.346483017" Oct 02 12:53:29 crc kubenswrapper[4710]: I1002 12:53:29.279268 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8scqr" podStartSLOduration=2.8457632999999998 podStartE2EDuration="4.279250174s" podCreationTimestamp="2025-10-02 12:53:25 +0000 UTC" firstStartedPulling="2025-10-02 12:53:27.169346292 +0000 UTC m=+251.275757175" lastFinishedPulling="2025-10-02 12:53:28.602833166 +0000 UTC m=+252.709244049" observedRunningTime="2025-10-02 12:53:29.277205335 +0000 UTC m=+253.383616228" watchObservedRunningTime="2025-10-02 12:53:29.279250174 +0000 UTC m=+253.385661047" Oct 02 12:53:32 crc kubenswrapper[4710]: I1002 12:53:32.232162 4710 generic.go:334] "Generic (PLEG): container finished" podID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerID="1832a4f418c943fe425198f149841158cc423af7792857c87b80fb112ce945d2" exitCode=0 Oct 02 12:53:32 crc kubenswrapper[4710]: I1002 12:53:32.232240 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm6bc" event={"ID":"fccaa30c-236a-4f2a-a2de-644399d3679d","Type":"ContainerDied","Data":"1832a4f418c943fe425198f149841158cc423af7792857c87b80fb112ce945d2"} Oct 02 12:53:32 crc kubenswrapper[4710]: I1002 12:53:32.238510 4710 generic.go:334] "Generic (PLEG): container finished" podID="87ad1040-7c7d-4265-b428-d95694c6c346" containerID="e993b744b169faac1349273241e15efb5acd31e92e00b03f11aa186d578c82d6" exitCode=0 Oct 02 12:53:32 crc kubenswrapper[4710]: I1002 12:53:32.238569 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqkkb" event={"ID":"87ad1040-7c7d-4265-b428-d95694c6c346","Type":"ContainerDied","Data":"e993b744b169faac1349273241e15efb5acd31e92e00b03f11aa186d578c82d6"} Oct 02 12:53:33 crc kubenswrapper[4710]: I1002 12:53:33.245635 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm6bc" event={"ID":"fccaa30c-236a-4f2a-a2de-644399d3679d","Type":"ContainerStarted","Data":"e6690db660b4206d7defb2f2f64f4bf7739773f1549a8b3dcb66cd6070666d20"} Oct 02 12:53:33 crc kubenswrapper[4710]: I1002 12:53:33.248295 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqkkb" event={"ID":"87ad1040-7c7d-4265-b428-d95694c6c346","Type":"ContainerStarted","Data":"dffefae84ebb736a531c2f6bb6b931895620e84933566e0f8df85e4af134894e"} Oct 02 12:53:33 crc kubenswrapper[4710]: I1002 12:53:33.264614 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tm6bc" podStartSLOduration=1.503305723 podStartE2EDuration="5.264596769s" podCreationTimestamp="2025-10-02 12:53:28 +0000 UTC" firstStartedPulling="2025-10-02 12:53:29.212583715 +0000 UTC m=+253.318994598" lastFinishedPulling="2025-10-02 12:53:32.973874771 +0000 UTC m=+257.080285644" observedRunningTime="2025-10-02 12:53:33.262177969 +0000 UTC m=+257.368588852" watchObservedRunningTime="2025-10-02 12:53:33.264596769 +0000 UTC m=+257.371007652" Oct 02 12:53:33 crc kubenswrapper[4710]: I1002 12:53:33.282299 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hqkkb" podStartSLOduration=2.847028272 podStartE2EDuration="6.282280884s" podCreationTimestamp="2025-10-02 12:53:27 +0000 UTC" firstStartedPulling="2025-10-02 12:53:29.202991605 +0000 UTC m=+253.309402488" lastFinishedPulling="2025-10-02 12:53:32.638244207 +0000 UTC m=+256.744655100" observedRunningTime="2025-10-02 12:53:33.280413449 +0000 UTC m=+257.386824332" watchObservedRunningTime="2025-10-02 12:53:33.282280884 +0000 UTC m=+257.388691767" Oct 02 12:53:35 crc kubenswrapper[4710]: I1002 12:53:35.900993 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:35 crc kubenswrapper[4710]: I1002 12:53:35.901389 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:35 crc kubenswrapper[4710]: I1002 12:53:35.945170 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:36 crc kubenswrapper[4710]: I1002 12:53:36.132170 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:36 crc kubenswrapper[4710]: I1002 12:53:36.132246 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:36 crc kubenswrapper[4710]: I1002 12:53:36.177258 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:36 crc kubenswrapper[4710]: I1002 12:53:36.316221 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lb6tp" Oct 02 12:53:36 crc kubenswrapper[4710]: I1002 12:53:36.319465 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8scqr" Oct 02 12:53:38 crc kubenswrapper[4710]: I1002 12:53:38.287590 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:38 crc kubenswrapper[4710]: I1002 12:53:38.288496 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:38 crc kubenswrapper[4710]: I1002 12:53:38.322600 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:38 crc kubenswrapper[4710]: I1002 12:53:38.581486 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:38 crc kubenswrapper[4710]: I1002 12:53:38.581530 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:38 crc kubenswrapper[4710]: I1002 12:53:38.638610 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:53:39 crc kubenswrapper[4710]: I1002 12:53:39.324097 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hqkkb" Oct 02 12:53:39 crc kubenswrapper[4710]: I1002 12:53:39.325415 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 12:55:22 crc kubenswrapper[4710]: I1002 12:55:22.531081 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:55:22 crc kubenswrapper[4710]: I1002 12:55:22.531998 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.747660 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-94fd2"] Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.749270 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.770353 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-94fd2"] Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.849679 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/660d0a92-fb41-4ac9-a499-89f57302ca7a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.849726 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/660d0a92-fb41-4ac9-a499-89f57302ca7a-trusted-ca\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.849774 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn2lr\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-kube-api-access-dn2lr\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.849939 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.849985 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-bound-sa-token\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.850042 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/660d0a92-fb41-4ac9-a499-89f57302ca7a-registry-certificates\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.850082 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/660d0a92-fb41-4ac9-a499-89f57302ca7a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.850104 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-registry-tls\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.872110 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951087 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/660d0a92-fb41-4ac9-a499-89f57302ca7a-trusted-ca\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951174 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn2lr\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-kube-api-access-dn2lr\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951224 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-bound-sa-token\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951262 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/660d0a92-fb41-4ac9-a499-89f57302ca7a-registry-certificates\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951296 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/660d0a92-fb41-4ac9-a499-89f57302ca7a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951321 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-registry-tls\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.951355 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/660d0a92-fb41-4ac9-a499-89f57302ca7a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.952034 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/660d0a92-fb41-4ac9-a499-89f57302ca7a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.952440 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/660d0a92-fb41-4ac9-a499-89f57302ca7a-trusted-ca\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.952741 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/660d0a92-fb41-4ac9-a499-89f57302ca7a-registry-certificates\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.956980 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/660d0a92-fb41-4ac9-a499-89f57302ca7a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.957008 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-registry-tls\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.967170 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-bound-sa-token\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:36 crc kubenswrapper[4710]: I1002 12:55:36.968277 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn2lr\" (UniqueName: \"kubernetes.io/projected/660d0a92-fb41-4ac9-a499-89f57302ca7a-kube-api-access-dn2lr\") pod \"image-registry-66df7c8f76-94fd2\" (UID: \"660d0a92-fb41-4ac9-a499-89f57302ca7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:37 crc kubenswrapper[4710]: I1002 12:55:37.067101 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:37 crc kubenswrapper[4710]: I1002 12:55:37.237996 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-94fd2"] Oct 02 12:55:38 crc kubenswrapper[4710]: I1002 12:55:38.006931 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" event={"ID":"660d0a92-fb41-4ac9-a499-89f57302ca7a","Type":"ContainerStarted","Data":"a71a731906d7619f9de8abac41048a1e8e8a4a84228e8494d27d0e1a73e9bcba"} Oct 02 12:55:38 crc kubenswrapper[4710]: I1002 12:55:38.007280 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" event={"ID":"660d0a92-fb41-4ac9-a499-89f57302ca7a","Type":"ContainerStarted","Data":"31d6aa9d4500a96dbb11641e37e0776cede37af5ca10f1aed187a7e675350d86"} Oct 02 12:55:38 crc kubenswrapper[4710]: I1002 12:55:38.007303 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:38 crc kubenswrapper[4710]: I1002 12:55:38.025695 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" podStartSLOduration=2.025671777 podStartE2EDuration="2.025671777s" podCreationTimestamp="2025-10-02 12:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:55:38.023446744 +0000 UTC m=+382.129857637" watchObservedRunningTime="2025-10-02 12:55:38.025671777 +0000 UTC m=+382.132082670" Oct 02 12:55:52 crc kubenswrapper[4710]: I1002 12:55:52.530727 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:55:52 crc kubenswrapper[4710]: I1002 12:55:52.531169 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:55:57 crc kubenswrapper[4710]: I1002 12:55:57.078046 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-94fd2" Oct 02 12:55:57 crc kubenswrapper[4710]: I1002 12:55:57.153299 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cx6br"] Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.220832 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" podUID="db597145-79db-4f18-ae0c-c64f35af2fcc" containerName="registry" containerID="cri-o://478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019" gracePeriod=30 Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.530087 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.530440 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.530481 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.531148 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1f26514b5ee22a80a5d66c96251f5d904520e9599cd1deeb21c078bb6cf25a5"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.532157 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://d1f26514b5ee22a80a5d66c96251f5d904520e9599cd1deeb21c078bb6cf25a5" gracePeriod=600 Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.560004 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.680673 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.680711 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qj77\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-kube-api-access-4qj77\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.680809 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db597145-79db-4f18-ae0c-c64f35af2fcc-ca-trust-extracted\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.680888 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-certificates\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.680916 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-tls\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.681681 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-trusted-ca\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.681692 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.681721 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-bound-sa-token\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.681785 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db597145-79db-4f18-ae0c-c64f35af2fcc-installation-pull-secrets\") pod \"db597145-79db-4f18-ae0c-c64f35af2fcc\" (UID: \"db597145-79db-4f18-ae0c-c64f35af2fcc\") " Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.682032 4710 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.682375 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.685862 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.686636 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db597145-79db-4f18-ae0c-c64f35af2fcc-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.689733 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-kube-api-access-4qj77" (OuterVolumeSpecName: "kube-api-access-4qj77") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "kube-api-access-4qj77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.690228 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.690474 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.701205 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db597145-79db-4f18-ae0c-c64f35af2fcc-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "db597145-79db-4f18-ae0c-c64f35af2fcc" (UID: "db597145-79db-4f18-ae0c-c64f35af2fcc"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.783284 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db597145-79db-4f18-ae0c-c64f35af2fcc-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.783318 4710 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.783330 4710 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db597145-79db-4f18-ae0c-c64f35af2fcc-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.783339 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qj77\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-kube-api-access-4qj77\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.783347 4710 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db597145-79db-4f18-ae0c-c64f35af2fcc-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:22 crc kubenswrapper[4710]: I1002 12:56:22.783355 4710 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db597145-79db-4f18-ae0c-c64f35af2fcc-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.272730 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="d1f26514b5ee22a80a5d66c96251f5d904520e9599cd1deeb21c078bb6cf25a5" exitCode=0 Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.272821 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"d1f26514b5ee22a80a5d66c96251f5d904520e9599cd1deeb21c078bb6cf25a5"} Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.273124 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"579103ebcb45d750aaacc6f76838d855e05a1b531da1b682b820b0d029b5b04d"} Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.273148 4710 scope.go:117] "RemoveContainer" containerID="c74897770a2756676f97580d869ba8cb990d5f50d1d86308606a3181441ecc9c" Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.275189 4710 generic.go:334] "Generic (PLEG): container finished" podID="db597145-79db-4f18-ae0c-c64f35af2fcc" containerID="478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019" exitCode=0 Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.275222 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" event={"ID":"db597145-79db-4f18-ae0c-c64f35af2fcc","Type":"ContainerDied","Data":"478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019"} Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.275242 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.275248 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cx6br" event={"ID":"db597145-79db-4f18-ae0c-c64f35af2fcc","Type":"ContainerDied","Data":"0a2ad9d40366a0005d408a7f3616925d8c6ff199fb192e9a35ca1c4bd79d2f77"} Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.295204 4710 scope.go:117] "RemoveContainer" containerID="478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019" Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.304045 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cx6br"] Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.310192 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cx6br"] Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.313101 4710 scope.go:117] "RemoveContainer" containerID="478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019" Oct 02 12:56:23 crc kubenswrapper[4710]: E1002 12:56:23.313650 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019\": container with ID starting with 478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019 not found: ID does not exist" containerID="478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019" Oct 02 12:56:23 crc kubenswrapper[4710]: I1002 12:56:23.313680 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019"} err="failed to get container status \"478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019\": rpc error: code = NotFound desc = could not find container \"478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019\": container with ID starting with 478b6421db860054b9369ea0f0422003d63440c955ab6cae6c0de10446faf019 not found: ID does not exist" Oct 02 12:56:24 crc kubenswrapper[4710]: I1002 12:56:24.915910 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db597145-79db-4f18-ae0c-c64f35af2fcc" path="/var/lib/kubelet/pods/db597145-79db-4f18-ae0c-c64f35af2fcc/volumes" Oct 02 12:58:22 crc kubenswrapper[4710]: I1002 12:58:22.530583 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:58:22 crc kubenswrapper[4710]: I1002 12:58:22.531243 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:58:52 crc kubenswrapper[4710]: I1002 12:58:52.530263 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:58:52 crc kubenswrapper[4710]: I1002 12:58:52.530665 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.667985 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wbjnf"] Oct 02 12:59:06 crc kubenswrapper[4710]: E1002 12:59:06.668526 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db597145-79db-4f18-ae0c-c64f35af2fcc" containerName="registry" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.668541 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="db597145-79db-4f18-ae0c-c64f35af2fcc" containerName="registry" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.668667 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="db597145-79db-4f18-ae0c-c64f35af2fcc" containerName="registry" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.669144 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.671033 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.671678 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.674492 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9p576"] Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.675414 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9p576" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.676900 4710 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-mv2rp" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.679644 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wbjnf"] Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.691908 4710 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-2qlv4" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.701430 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rsnsq"] Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.702215 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.708190 4710 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-tsnhc" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.709665 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9p576"] Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.712501 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htnsw\" (UniqueName: \"kubernetes.io/projected/4d218b44-c736-49fa-b12a-0c21a5ca361b-kube-api-access-htnsw\") pod \"cert-manager-cainjector-7f985d654d-wbjnf\" (UID: \"4d218b44-c736-49fa-b12a-0c21a5ca361b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.712681 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fqrk\" (UniqueName: \"kubernetes.io/projected/7181be4b-6c46-46e5-8d15-453f1b81ed21-kube-api-access-7fqrk\") pod \"cert-manager-webhook-5655c58dd6-rsnsq\" (UID: \"7181be4b-6c46-46e5-8d15-453f1b81ed21\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.712801 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfzf8\" (UniqueName: \"kubernetes.io/projected/4f8b7f42-0f95-448d-963c-9cfce5078d2b-kube-api-access-cfzf8\") pod \"cert-manager-5b446d88c5-9p576\" (UID: \"4f8b7f42-0f95-448d-963c-9cfce5078d2b\") " pod="cert-manager/cert-manager-5b446d88c5-9p576" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.715376 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rsnsq"] Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.814131 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fqrk\" (UniqueName: \"kubernetes.io/projected/7181be4b-6c46-46e5-8d15-453f1b81ed21-kube-api-access-7fqrk\") pod \"cert-manager-webhook-5655c58dd6-rsnsq\" (UID: \"7181be4b-6c46-46e5-8d15-453f1b81ed21\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.814206 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfzf8\" (UniqueName: \"kubernetes.io/projected/4f8b7f42-0f95-448d-963c-9cfce5078d2b-kube-api-access-cfzf8\") pod \"cert-manager-5b446d88c5-9p576\" (UID: \"4f8b7f42-0f95-448d-963c-9cfce5078d2b\") " pod="cert-manager/cert-manager-5b446d88c5-9p576" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.814272 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htnsw\" (UniqueName: \"kubernetes.io/projected/4d218b44-c736-49fa-b12a-0c21a5ca361b-kube-api-access-htnsw\") pod \"cert-manager-cainjector-7f985d654d-wbjnf\" (UID: \"4d218b44-c736-49fa-b12a-0c21a5ca361b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.832962 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htnsw\" (UniqueName: \"kubernetes.io/projected/4d218b44-c736-49fa-b12a-0c21a5ca361b-kube-api-access-htnsw\") pod \"cert-manager-cainjector-7f985d654d-wbjnf\" (UID: \"4d218b44-c736-49fa-b12a-0c21a5ca361b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.836303 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfzf8\" (UniqueName: \"kubernetes.io/projected/4f8b7f42-0f95-448d-963c-9cfce5078d2b-kube-api-access-cfzf8\") pod \"cert-manager-5b446d88c5-9p576\" (UID: \"4f8b7f42-0f95-448d-963c-9cfce5078d2b\") " pod="cert-manager/cert-manager-5b446d88c5-9p576" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.837399 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fqrk\" (UniqueName: \"kubernetes.io/projected/7181be4b-6c46-46e5-8d15-453f1b81ed21-kube-api-access-7fqrk\") pod \"cert-manager-webhook-5655c58dd6-rsnsq\" (UID: \"7181be4b-6c46-46e5-8d15-453f1b81ed21\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.994838 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" Oct 02 12:59:06 crc kubenswrapper[4710]: I1002 12:59:06.998016 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9p576" Oct 02 12:59:07 crc kubenswrapper[4710]: I1002 12:59:07.021294 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:07 crc kubenswrapper[4710]: I1002 12:59:07.210867 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wbjnf"] Oct 02 12:59:07 crc kubenswrapper[4710]: I1002 12:59:07.217956 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 12:59:07 crc kubenswrapper[4710]: I1002 12:59:07.273930 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" event={"ID":"4d218b44-c736-49fa-b12a-0c21a5ca361b","Type":"ContainerStarted","Data":"71fcd6b50a9a9270d30da5a46c41efcd3f252caa4dc60cd9141f56d7047385e0"} Oct 02 12:59:07 crc kubenswrapper[4710]: I1002 12:59:07.503098 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9p576"] Oct 02 12:59:07 crc kubenswrapper[4710]: I1002 12:59:07.510303 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rsnsq"] Oct 02 12:59:07 crc kubenswrapper[4710]: W1002 12:59:07.511331 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f8b7f42_0f95_448d_963c_9cfce5078d2b.slice/crio-9b8ada27b7a71a9efbbc0b5a295c0cb9c02e4858ecc2e8fd83e61a3551505451 WatchSource:0}: Error finding container 9b8ada27b7a71a9efbbc0b5a295c0cb9c02e4858ecc2e8fd83e61a3551505451: Status 404 returned error can't find the container with id 9b8ada27b7a71a9efbbc0b5a295c0cb9c02e4858ecc2e8fd83e61a3551505451 Oct 02 12:59:08 crc kubenswrapper[4710]: I1002 12:59:08.279816 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" event={"ID":"7181be4b-6c46-46e5-8d15-453f1b81ed21","Type":"ContainerStarted","Data":"e98a26c7bfb782f79cd408045dfa342dae1d536a86e7c4a090c6bf2908661bb3"} Oct 02 12:59:08 crc kubenswrapper[4710]: I1002 12:59:08.281231 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9p576" event={"ID":"4f8b7f42-0f95-448d-963c-9cfce5078d2b","Type":"ContainerStarted","Data":"9b8ada27b7a71a9efbbc0b5a295c0cb9c02e4858ecc2e8fd83e61a3551505451"} Oct 02 12:59:10 crc kubenswrapper[4710]: I1002 12:59:10.318030 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" event={"ID":"4d218b44-c736-49fa-b12a-0c21a5ca361b","Type":"ContainerStarted","Data":"13350e63981595aeeb59702a2e39eb823a2b240a2bbfdab5751dfe3aef29475a"} Oct 02 12:59:12 crc kubenswrapper[4710]: I1002 12:59:12.328954 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" event={"ID":"7181be4b-6c46-46e5-8d15-453f1b81ed21","Type":"ContainerStarted","Data":"e3bba72ab6b1c14ddd9646d1d98271b51aef6f264cacdc4026d29415734539f5"} Oct 02 12:59:12 crc kubenswrapper[4710]: I1002 12:59:12.329111 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:12 crc kubenswrapper[4710]: I1002 12:59:12.331206 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9p576" event={"ID":"4f8b7f42-0f95-448d-963c-9cfce5078d2b","Type":"ContainerStarted","Data":"5d5e9310c569d45f22ab7d2bf2dcaab86259aa59112a57f61a88e424d89b4976"} Oct 02 12:59:12 crc kubenswrapper[4710]: I1002 12:59:12.345849 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-wbjnf" podStartSLOduration=4.398813399 podStartE2EDuration="6.345829978s" podCreationTimestamp="2025-10-02 12:59:06 +0000 UTC" firstStartedPulling="2025-10-02 12:59:07.217768289 +0000 UTC m=+591.324179172" lastFinishedPulling="2025-10-02 12:59:09.164784858 +0000 UTC m=+593.271195751" observedRunningTime="2025-10-02 12:59:10.347619516 +0000 UTC m=+594.454030399" watchObservedRunningTime="2025-10-02 12:59:12.345829978 +0000 UTC m=+596.452240871" Oct 02 12:59:12 crc kubenswrapper[4710]: I1002 12:59:12.347036 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" podStartSLOduration=2.331784977 podStartE2EDuration="6.347029165s" podCreationTimestamp="2025-10-02 12:59:06 +0000 UTC" firstStartedPulling="2025-10-02 12:59:07.530255567 +0000 UTC m=+591.636666460" lastFinishedPulling="2025-10-02 12:59:11.545499765 +0000 UTC m=+595.651910648" observedRunningTime="2025-10-02 12:59:12.341158432 +0000 UTC m=+596.447569305" watchObservedRunningTime="2025-10-02 12:59:12.347029165 +0000 UTC m=+596.453440048" Oct 02 12:59:12 crc kubenswrapper[4710]: I1002 12:59:12.357731 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-9p576" podStartSLOduration=2.331737265 podStartE2EDuration="6.357705847s" podCreationTimestamp="2025-10-02 12:59:06 +0000 UTC" firstStartedPulling="2025-10-02 12:59:07.514915959 +0000 UTC m=+591.621326852" lastFinishedPulling="2025-10-02 12:59:11.540884551 +0000 UTC m=+595.647295434" observedRunningTime="2025-10-02 12:59:12.352809016 +0000 UTC m=+596.459219979" watchObservedRunningTime="2025-10-02 12:59:12.357705847 +0000 UTC m=+596.464116770" Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.814697 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8jhhq"] Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.815715 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-controller" containerID="cri-o://43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.815808 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.815825 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="nbdb" containerID="cri-o://b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.815899 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="northd" containerID="cri-o://b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.815885 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-node" containerID="cri-o://03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.815873 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-acl-logging" containerID="cri-o://06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.816063 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="sbdb" containerID="cri-o://b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37" gracePeriod=30 Oct 02 12:59:16 crc kubenswrapper[4710]: I1002 12:59:16.862016 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" containerID="cri-o://0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a" gracePeriod=30 Oct 02 12:59:17 crc kubenswrapper[4710]: I1002 12:59:17.024713 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-rsnsq" Oct 02 12:59:17 crc kubenswrapper[4710]: I1002 12:59:17.361321 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/2.log" Oct 02 12:59:17 crc kubenswrapper[4710]: I1002 12:59:17.363155 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-acl-logging/0.log" Oct 02 12:59:17 crc kubenswrapper[4710]: I1002 12:59:17.363712 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9" exitCode=143 Oct 02 12:59:17 crc kubenswrapper[4710]: I1002 12:59:17.363757 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.371838 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8ddjb_7786a93b-a392-425e-b4ac-ed4956546d08/kube-multus/1.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.372482 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8ddjb_7786a93b-a392-425e-b4ac-ed4956546d08/kube-multus/0.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.372578 4710 generic.go:334] "Generic (PLEG): container finished" podID="7786a93b-a392-425e-b4ac-ed4956546d08" containerID="581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9" exitCode=2 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.372650 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8ddjb" event={"ID":"7786a93b-a392-425e-b4ac-ed4956546d08","Type":"ContainerDied","Data":"581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.372699 4710 scope.go:117] "RemoveContainer" containerID="b0b9ca6b66ddb0e150b260aadda8f8290d242092e30052092b05d2d3c1c9508d" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.373548 4710 scope.go:117] "RemoveContainer" containerID="581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.374007 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8ddjb_openshift-multus(7786a93b-a392-425e-b4ac-ed4956546d08)\"" pod="openshift-multus/multus-8ddjb" podUID="7786a93b-a392-425e-b4ac-ed4956546d08" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.377045 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/2.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.382997 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-acl-logging/0.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384108 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-controller/0.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384801 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a" exitCode=0 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384839 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37" exitCode=0 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384857 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70" exitCode=0 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384866 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d" exitCode=0 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384877 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e" exitCode=0 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384889 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81" exitCode=0 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384902 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerID="43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f" exitCode=143 Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384903 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384945 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384960 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384975 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.384989 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.385001 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.385014 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f"} Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.488061 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/2.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.490848 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-acl-logging/0.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.491514 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-controller/0.log" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.492058 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553363 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mpbf8"] Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553585 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="nbdb" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553602 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="nbdb" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553614 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="northd" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553622 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="northd" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553635 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553645 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553655 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553662 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553673 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-node" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553681 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-node" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553693 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="sbdb" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553703 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="sbdb" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553713 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-acl-logging" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553721 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-acl-logging" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553732 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kubecfg-setup" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553738 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kubecfg-setup" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553774 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553782 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.553793 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553801 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553906 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="sbdb" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553919 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553930 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-node" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553943 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553953 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="nbdb" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553961 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-acl-logging" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553970 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553978 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovn-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553988 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.553996 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.554004 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="northd" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.554115 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.554126 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: E1002 12:59:18.554139 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.554150 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" containerName="ovnkube-controller" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.556584 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583427 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-var-lib-openvswitch\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583483 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlht4\" (UniqueName: \"kubernetes.io/projected/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-kube-api-access-wlht4\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583510 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-netns\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583543 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-env-overrides\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583567 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-systemd-units\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583587 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-log-socket\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583614 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-netd\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583642 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583665 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-bin\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583694 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-slash\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583770 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-systemd\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583798 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-config\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583848 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovn-node-metrics-cert\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583896 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-script-lib\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583918 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-ovn-kubernetes\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583943 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-kubelet\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583967 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-etc-openvswitch\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.583989 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-ovn\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584007 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-openvswitch\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584025 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-node-log\") pod \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\" (UID: \"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12\") " Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584171 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-run-netns\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584201 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-systemd-units\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584223 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-ovnkube-config\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584240 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584282 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584310 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584333 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-node-log" (OuterVolumeSpecName: "node-log") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584421 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584707 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584734 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584741 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584799 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584824 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584849 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584870 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-log-socket" (OuterVolumeSpecName: "log-socket") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584890 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584912 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-slash" (OuterVolumeSpecName: "host-slash") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584934 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584251 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-systemd\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.584983 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-ovnkube-script-lib\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585009 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-run-ovn-kubernetes\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585034 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-cni-bin\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585056 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-log-socket\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585062 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585074 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4qv\" (UniqueName: \"kubernetes.io/projected/5990f341-3b74-4756-9622-4741a9f32474-kube-api-access-6l4qv\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585102 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-var-lib-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585143 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585164 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585186 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-kubelet\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585218 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-etc-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585239 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-slash\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585258 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-cni-netd\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585278 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5990f341-3b74-4756-9622-4741a9f32474-ovn-node-metrics-cert\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585310 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-node-log\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585341 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-env-overrides\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585364 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-ovn\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585406 4710 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585419 4710 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585431 4710 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585443 4710 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585455 4710 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585467 4710 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585478 4710 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585489 4710 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585500 4710 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585512 4710 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585523 4710 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585535 4710 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585544 4710 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585556 4710 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585567 4710 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585578 4710 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.585590 4710 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.589840 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.594887 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-kube-api-access-wlht4" (OuterVolumeSpecName: "kube-api-access-wlht4") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "kube-api-access-wlht4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.609421 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" (UID: "93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.662829 4710 scope.go:117] "RemoveContainer" containerID="e1de915b5b3e5da53f3acbdf70fc663dc354d7f9719b07ca0f937d25b9eaba6a" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687218 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-run-netns\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687263 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-systemd-units\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687286 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-ovnkube-config\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687309 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-systemd\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687337 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-ovnkube-script-lib\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687354 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-run-netns\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687366 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-systemd-units\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687399 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-run-ovn-kubernetes\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687356 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-run-ovn-kubernetes\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687448 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-cni-bin\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687464 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-log-socket\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687481 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4qv\" (UniqueName: \"kubernetes.io/projected/5990f341-3b74-4756-9622-4741a9f32474-kube-api-access-6l4qv\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687506 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-var-lib-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687524 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687544 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687540 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-systemd\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687585 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-kubelet\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687588 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-var-lib-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687561 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-kubelet\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687619 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687557 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-log-socket\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687624 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-cni-bin\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687730 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687855 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-etc-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687911 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-slash\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687946 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5990f341-3b74-4756-9622-4741a9f32474-ovn-node-metrics-cert\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687967 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-slash\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687982 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-cni-netd\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688047 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-node-log\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688089 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-ovnkube-config\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688108 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-env-overrides\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688128 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-host-cni-netd\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.687948 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-etc-openvswitch\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688161 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-node-log\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688202 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-ovn\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688224 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-ovnkube-script-lib\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688329 4710 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688354 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlht4\" (UniqueName: \"kubernetes.io/projected/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-kube-api-access-wlht4\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688366 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5990f341-3b74-4756-9622-4741a9f32474-run-ovn\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688374 4710 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.688467 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5990f341-3b74-4756-9622-4741a9f32474-env-overrides\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.691136 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5990f341-3b74-4756-9622-4741a9f32474-ovn-node-metrics-cert\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.702053 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4qv\" (UniqueName: \"kubernetes.io/projected/5990f341-3b74-4756-9622-4741a9f32474-kube-api-access-6l4qv\") pod \"ovnkube-node-mpbf8\" (UID: \"5990f341-3b74-4756-9622-4741a9f32474\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: I1002 12:59:18.870038 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:18 crc kubenswrapper[4710]: W1002 12:59:18.891202 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5990f341_3b74_4756_9622_4741a9f32474.slice/crio-5bff479e3a613a96f600d3bdfd9cb270320f748f19e250da61b08e60f526f873 WatchSource:0}: Error finding container 5bff479e3a613a96f600d3bdfd9cb270320f748f19e250da61b08e60f526f873: Status 404 returned error can't find the container with id 5bff479e3a613a96f600d3bdfd9cb270320f748f19e250da61b08e60f526f873 Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.398323 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-acl-logging/0.log" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.399475 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovn-controller/0.log" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.400105 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" event={"ID":"93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12","Type":"ContainerDied","Data":"26968094c576b9eb7c131a4671b065d80104a9deef4b43b8bffe0bc631d967b8"} Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.400155 4710 scope.go:117] "RemoveContainer" containerID="0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.400395 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8jhhq" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.402300 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"5bff479e3a613a96f600d3bdfd9cb270320f748f19e250da61b08e60f526f873"} Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.408838 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8ddjb_7786a93b-a392-425e-b4ac-ed4956546d08/kube-multus/1.log" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.443802 4710 scope.go:117] "RemoveContainer" containerID="b9ec71374dabeab65c3482d7bd5461322442223cb1e00fd11ac2423b07048d37" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.475229 4710 scope.go:117] "RemoveContainer" containerID="b1ef687a66f0a4b5cb7849b278859bbc2b8cbbd90fca7f1db3224ab3394d9f70" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.498794 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8jhhq"] Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.514623 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8jhhq"] Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.521292 4710 scope.go:117] "RemoveContainer" containerID="b88d5951f5e0a4c7400d53437e6ce91782e38b4244797d67577fbf4cdf9be60d" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.573117 4710 scope.go:117] "RemoveContainer" containerID="873f338be8407527be3c1fde61e2bf803ccb819430aefa71f5118c5da218175e" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.591869 4710 scope.go:117] "RemoveContainer" containerID="03fe61ab48889dfa5ff0de7a4286369f5c457f2c47ee38bac81ab4fc9ddbfc81" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.606391 4710 scope.go:117] "RemoveContainer" containerID="06a4e947823dd08c9a6f71f8376cdd74001aeba02aafe0b71d48f4c3599d9ef9" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.618399 4710 scope.go:117] "RemoveContainer" containerID="43c786bb63ff5e18c0a1b5ae9c654159e2ccd9e2058defe2d950399e6ceff69f" Oct 02 12:59:19 crc kubenswrapper[4710]: I1002 12:59:19.631089 4710 scope.go:117] "RemoveContainer" containerID="caec85cd3fc958d8d9d98fb2ef294c613252bdbcc7de3dd454df6f0743c81db1" Oct 02 12:59:20 crc kubenswrapper[4710]: I1002 12:59:20.416556 4710 generic.go:334] "Generic (PLEG): container finished" podID="5990f341-3b74-4756-9622-4741a9f32474" containerID="2567cd18111aedb13685ea9e2b7812c77852544f5fea4df4c012212ec22d01c5" exitCode=0 Oct 02 12:59:20 crc kubenswrapper[4710]: I1002 12:59:20.416594 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerDied","Data":"2567cd18111aedb13685ea9e2b7812c77852544f5fea4df4c012212ec22d01c5"} Oct 02 12:59:20 crc kubenswrapper[4710]: I1002 12:59:20.913891 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12" path="/var/lib/kubelet/pods/93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/volumes" Oct 02 12:59:21 crc kubenswrapper[4710]: I1002 12:59:21.425921 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"cb1f2b0eb9ec966fddf41cfb7ca817e20205e98933ee91b12b95c9679e01e66d"} Oct 02 12:59:21 crc kubenswrapper[4710]: I1002 12:59:21.426008 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"d78b2997d88441b2d8fce36ffc627bf7e10ce7bc65a6c066d4d5da6e376f2c23"} Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.436505 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"5931ee6235e9d90a873a733fc4fc5d387c3549508dc204ac9e9387abd3874446"} Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.436848 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"5fc9895bcfecd7616906bd9211d0cb704ab434e82aa254a3fa31cdcafc1809f2"} Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.533290 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.533359 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.533410 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.534080 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"579103ebcb45d750aaacc6f76838d855e05a1b531da1b682b820b0d029b5b04d"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:59:22 crc kubenswrapper[4710]: I1002 12:59:22.534167 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://579103ebcb45d750aaacc6f76838d855e05a1b531da1b682b820b0d029b5b04d" gracePeriod=600 Oct 02 12:59:23 crc kubenswrapper[4710]: I1002 12:59:23.445517 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="579103ebcb45d750aaacc6f76838d855e05a1b531da1b682b820b0d029b5b04d" exitCode=0 Oct 02 12:59:23 crc kubenswrapper[4710]: I1002 12:59:23.445588 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"579103ebcb45d750aaacc6f76838d855e05a1b531da1b682b820b0d029b5b04d"} Oct 02 12:59:23 crc kubenswrapper[4710]: I1002 12:59:23.445968 4710 scope.go:117] "RemoveContainer" containerID="d1f26514b5ee22a80a5d66c96251f5d904520e9599cd1deeb21c078bb6cf25a5" Oct 02 12:59:23 crc kubenswrapper[4710]: I1002 12:59:23.455178 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"28253ca805ac1252835072b7a2063024a4d507e9bc8adebe30576590d575eab7"} Oct 02 12:59:23 crc kubenswrapper[4710]: I1002 12:59:23.455234 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"a246e16494a2826846853f294cf3e84b12e3f8146701176da6e2db9696f150cc"} Oct 02 12:59:24 crc kubenswrapper[4710]: E1002 12:59:24.168940 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 12:59:24 crc kubenswrapper[4710]: I1002 12:59:24.464565 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"2554f6884f6fea884aeb14ab54b4450d5c8dcfe9721cc87320ad6736f5b0d86e"} Oct 02 12:59:26 crc kubenswrapper[4710]: I1002 12:59:26.483976 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"e635eb5c1b2b7677b52ef9193e7a0779546c64aeb6ae8fed7773115ac20d79ed"} Oct 02 12:59:30 crc kubenswrapper[4710]: I1002 12:59:30.511307 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" event={"ID":"5990f341-3b74-4756-9622-4741a9f32474","Type":"ContainerStarted","Data":"566a43913e4f4c9523bb2dbc85505ff0906de675f5afb5a320c5f47638a1f399"} Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.517718 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.518187 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.518209 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.550199 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" podStartSLOduration=13.550185211 podStartE2EDuration="13.550185211s" podCreationTimestamp="2025-10-02 12:59:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:59:31.548621915 +0000 UTC m=+615.655032918" watchObservedRunningTime="2025-10-02 12:59:31.550185211 +0000 UTC m=+615.656596084" Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.559234 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.561780 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:31 crc kubenswrapper[4710]: I1002 12:59:31.904843 4710 scope.go:117] "RemoveContainer" containerID="581e767c40e6764dc472e345331af14359e14bbf4122dbe7e627ac405ea691b9" Oct 02 12:59:32 crc kubenswrapper[4710]: I1002 12:59:32.526684 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8ddjb_7786a93b-a392-425e-b4ac-ed4956546d08/kube-multus/1.log" Oct 02 12:59:32 crc kubenswrapper[4710]: I1002 12:59:32.527168 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8ddjb" event={"ID":"7786a93b-a392-425e-b4ac-ed4956546d08","Type":"ContainerStarted","Data":"1a023bc6b3d0deb18b47302566c47ea1b7b1aaa971d983da0a792022f744dc4c"} Oct 02 12:59:34 crc kubenswrapper[4710]: E1002 12:59:34.318374 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 12:59:44 crc kubenswrapper[4710]: E1002 12:59:44.455129 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.576991 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k"] Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.577935 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.579921 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.586377 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k"] Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.736963 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.737135 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fstmc\" (UniqueName: \"kubernetes.io/projected/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-kube-api-access-fstmc\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.737215 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.838317 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.838391 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fstmc\" (UniqueName: \"kubernetes.io/projected/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-kube-api-access-fstmc\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.838429 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.838848 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.838930 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.856246 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fstmc\" (UniqueName: \"kubernetes.io/projected/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-kube-api-access-fstmc\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:45 crc kubenswrapper[4710]: I1002 12:59:45.894958 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:46 crc kubenswrapper[4710]: I1002 12:59:46.086806 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k"] Oct 02 12:59:46 crc kubenswrapper[4710]: W1002 12:59:46.090860 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cf57a3f_610d_4ee5_b2d9_ed967b7bfbde.slice/crio-8a67124b7bab93eebd98b48729b1e5b037fefcfc637b20f22a3109537a1e3a0c WatchSource:0}: Error finding container 8a67124b7bab93eebd98b48729b1e5b037fefcfc637b20f22a3109537a1e3a0c: Status 404 returned error can't find the container with id 8a67124b7bab93eebd98b48729b1e5b037fefcfc637b20f22a3109537a1e3a0c Oct 02 12:59:46 crc kubenswrapper[4710]: I1002 12:59:46.604254 4710 generic.go:334] "Generic (PLEG): container finished" podID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerID="299c80b60575e04c060b5a076921c9e5e730814a68c78d1362b8edda86554d5d" exitCode=0 Oct 02 12:59:46 crc kubenswrapper[4710]: I1002 12:59:46.604537 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" event={"ID":"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde","Type":"ContainerDied","Data":"299c80b60575e04c060b5a076921c9e5e730814a68c78d1362b8edda86554d5d"} Oct 02 12:59:46 crc kubenswrapper[4710]: I1002 12:59:46.604570 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" event={"ID":"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde","Type":"ContainerStarted","Data":"8a67124b7bab93eebd98b48729b1e5b037fefcfc637b20f22a3109537a1e3a0c"} Oct 02 12:59:48 crc kubenswrapper[4710]: I1002 12:59:48.615345 4710 generic.go:334] "Generic (PLEG): container finished" podID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerID="cb0bfde710adb2e8ec069135bd60518a4d06ac1b823462c43e315c85601c8e63" exitCode=0 Oct 02 12:59:48 crc kubenswrapper[4710]: I1002 12:59:48.615410 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" event={"ID":"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde","Type":"ContainerDied","Data":"cb0bfde710adb2e8ec069135bd60518a4d06ac1b823462c43e315c85601c8e63"} Oct 02 12:59:48 crc kubenswrapper[4710]: I1002 12:59:48.896726 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mpbf8" Oct 02 12:59:49 crc kubenswrapper[4710]: I1002 12:59:49.623632 4710 generic.go:334] "Generic (PLEG): container finished" podID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerID="ff31763702a3402103a27c6e19e397e888f2c4157b11d16d8c5a3f6379a287c2" exitCode=0 Oct 02 12:59:49 crc kubenswrapper[4710]: I1002 12:59:49.623724 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" event={"ID":"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde","Type":"ContainerDied","Data":"ff31763702a3402103a27c6e19e397e888f2c4157b11d16d8c5a3f6379a287c2"} Oct 02 12:59:50 crc kubenswrapper[4710]: I1002 12:59:50.853333 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.003117 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-util\") pod \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.003217 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fstmc\" (UniqueName: \"kubernetes.io/projected/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-kube-api-access-fstmc\") pod \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.003296 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-bundle\") pod \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\" (UID: \"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde\") " Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.005322 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-bundle" (OuterVolumeSpecName: "bundle") pod "6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" (UID: "6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.008615 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-kube-api-access-fstmc" (OuterVolumeSpecName: "kube-api-access-fstmc") pod "6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" (UID: "6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde"). InnerVolumeSpecName "kube-api-access-fstmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.016135 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-util" (OuterVolumeSpecName: "util") pod "6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" (UID: "6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.104417 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fstmc\" (UniqueName: \"kubernetes.io/projected/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-kube-api-access-fstmc\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.104471 4710 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.104482 4710 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde-util\") on node \"crc\" DevicePath \"\"" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.650773 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" event={"ID":"6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde","Type":"ContainerDied","Data":"8a67124b7bab93eebd98b48729b1e5b037fefcfc637b20f22a3109537a1e3a0c"} Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.650850 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a67124b7bab93eebd98b48729b1e5b037fefcfc637b20f22a3109537a1e3a0c" Oct 02 12:59:51 crc kubenswrapper[4710]: I1002 12:59:51.650901 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k" Oct 02 12:59:54 crc kubenswrapper[4710]: E1002 12:59:54.564720 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.124339 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk"] Oct 02 13:00:00 crc kubenswrapper[4710]: E1002 13:00:00.125077 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="pull" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.125090 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="pull" Oct 02 13:00:00 crc kubenswrapper[4710]: E1002 13:00:00.125103 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="extract" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.125109 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="extract" Oct 02 13:00:00 crc kubenswrapper[4710]: E1002 13:00:00.125122 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="util" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.125128 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="util" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.125237 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde" containerName="extract" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.125584 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.129032 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.130218 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.151461 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk"] Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.222355 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96j75\" (UniqueName: \"kubernetes.io/projected/b0568be7-153e-4be2-adb6-605c17922339-kube-api-access-96j75\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.222431 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0568be7-153e-4be2-adb6-605c17922339-config-volume\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.222462 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0568be7-153e-4be2-adb6-605c17922339-secret-volume\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.323424 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96j75\" (UniqueName: \"kubernetes.io/projected/b0568be7-153e-4be2-adb6-605c17922339-kube-api-access-96j75\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.323502 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0568be7-153e-4be2-adb6-605c17922339-config-volume\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.323539 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0568be7-153e-4be2-adb6-605c17922339-secret-volume\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.324617 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0568be7-153e-4be2-adb6-605c17922339-config-volume\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.329181 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0568be7-153e-4be2-adb6-605c17922339-secret-volume\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.351002 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96j75\" (UniqueName: \"kubernetes.io/projected/b0568be7-153e-4be2-adb6-605c17922339-kube-api-access-96j75\") pod \"collect-profiles-29323500-tzbrk\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.443140 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:00 crc kubenswrapper[4710]: I1002 13:00:00.898099 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk"] Oct 02 13:00:01 crc kubenswrapper[4710]: I1002 13:00:01.708435 4710 generic.go:334] "Generic (PLEG): container finished" podID="b0568be7-153e-4be2-adb6-605c17922339" containerID="5743fc5ddd6c575120b5ee40cccf3796d479180679a6cb7edc08ba28568a33fc" exitCode=0 Oct 02 13:00:01 crc kubenswrapper[4710]: I1002 13:00:01.708520 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" event={"ID":"b0568be7-153e-4be2-adb6-605c17922339","Type":"ContainerDied","Data":"5743fc5ddd6c575120b5ee40cccf3796d479180679a6cb7edc08ba28568a33fc"} Oct 02 13:00:01 crc kubenswrapper[4710]: I1002 13:00:01.708640 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" event={"ID":"b0568be7-153e-4be2-adb6-605c17922339","Type":"ContainerStarted","Data":"f65f93b27bffcf4b41a048899673d15a7009fc3088c566c41e932ca46c939d09"} Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.658653 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.660351 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.667721 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.668137 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-sz4s9" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.668317 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.668454 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.703358 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.704211 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.705732 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-87bbn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.706176 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.714303 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.715021 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.722436 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.742809 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.754681 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blkr9\" (UniqueName: \"kubernetes.io/projected/4645af58-f9ab-4b88-a1fc-5a1ef95e330d-kube-api-access-blkr9\") pod \"obo-prometheus-operator-7c8cf85677-f8rs5\" (UID: \"4645af58-f9ab-4b88-a1fc-5a1ef95e330d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.855611 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d224b622-e6dd-47ce-868e-dfd2fe59d7c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2\" (UID: \"d224b622-e6dd-47ce-868e-dfd2fe59d7c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.855659 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3df7c825-06b5-480f-920e-47216a812bfd-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn\" (UID: \"3df7c825-06b5-480f-920e-47216a812bfd\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.855681 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3df7c825-06b5-480f-920e-47216a812bfd-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn\" (UID: \"3df7c825-06b5-480f-920e-47216a812bfd\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.855731 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blkr9\" (UniqueName: \"kubernetes.io/projected/4645af58-f9ab-4b88-a1fc-5a1ef95e330d-kube-api-access-blkr9\") pod \"obo-prometheus-operator-7c8cf85677-f8rs5\" (UID: \"4645af58-f9ab-4b88-a1fc-5a1ef95e330d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.855799 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d224b622-e6dd-47ce-868e-dfd2fe59d7c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2\" (UID: \"d224b622-e6dd-47ce-868e-dfd2fe59d7c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.888300 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-5jcdd"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.889201 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.893122 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-2hhf2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.893152 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.900627 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blkr9\" (UniqueName: \"kubernetes.io/projected/4645af58-f9ab-4b88-a1fc-5a1ef95e330d-kube-api-access-blkr9\") pod \"obo-prometheus-operator-7c8cf85677-f8rs5\" (UID: \"4645af58-f9ab-4b88-a1fc-5a1ef95e330d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.937348 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-5jcdd"] Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.956706 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d224b622-e6dd-47ce-868e-dfd2fe59d7c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2\" (UID: \"d224b622-e6dd-47ce-868e-dfd2fe59d7c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.956771 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3df7c825-06b5-480f-920e-47216a812bfd-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn\" (UID: \"3df7c825-06b5-480f-920e-47216a812bfd\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.956804 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3df7c825-06b5-480f-920e-47216a812bfd-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn\" (UID: \"3df7c825-06b5-480f-920e-47216a812bfd\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.956870 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d224b622-e6dd-47ce-868e-dfd2fe59d7c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2\" (UID: \"d224b622-e6dd-47ce-868e-dfd2fe59d7c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.959645 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d224b622-e6dd-47ce-868e-dfd2fe59d7c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2\" (UID: \"d224b622-e6dd-47ce-868e-dfd2fe59d7c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.960969 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3df7c825-06b5-480f-920e-47216a812bfd-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn\" (UID: \"3df7c825-06b5-480f-920e-47216a812bfd\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.962032 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d224b622-e6dd-47ce-868e-dfd2fe59d7c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2\" (UID: \"d224b622-e6dd-47ce-868e-dfd2fe59d7c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.962250 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3df7c825-06b5-480f-920e-47216a812bfd-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn\" (UID: \"3df7c825-06b5-480f-920e-47216a812bfd\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:02 crc kubenswrapper[4710]: I1002 13:00:02.978119 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.027107 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.036435 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.038000 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.058155 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbkn4\" (UniqueName: \"kubernetes.io/projected/20de3f43-1fd9-47e0-9fae-d84370233104-kube-api-access-gbkn4\") pod \"observability-operator-cc5f78dfc-5jcdd\" (UID: \"20de3f43-1fd9-47e0-9fae-d84370233104\") " pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.058246 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/20de3f43-1fd9-47e0-9fae-d84370233104-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-5jcdd\" (UID: \"20de3f43-1fd9-47e0-9fae-d84370233104\") " pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.086658 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-q42t7"] Oct 02 13:00:03 crc kubenswrapper[4710]: E1002 13:00:03.087096 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0568be7-153e-4be2-adb6-605c17922339" containerName="collect-profiles" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.087118 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0568be7-153e-4be2-adb6-605c17922339" containerName="collect-profiles" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.087252 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0568be7-153e-4be2-adb6-605c17922339" containerName="collect-profiles" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.087652 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.094317 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-kcgfm" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.105527 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-q42t7"] Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.159450 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96j75\" (UniqueName: \"kubernetes.io/projected/b0568be7-153e-4be2-adb6-605c17922339-kube-api-access-96j75\") pod \"b0568be7-153e-4be2-adb6-605c17922339\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.159607 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0568be7-153e-4be2-adb6-605c17922339-config-volume\") pod \"b0568be7-153e-4be2-adb6-605c17922339\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.159625 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0568be7-153e-4be2-adb6-605c17922339-secret-volume\") pod \"b0568be7-153e-4be2-adb6-605c17922339\" (UID: \"b0568be7-153e-4be2-adb6-605c17922339\") " Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.159797 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbkn4\" (UniqueName: \"kubernetes.io/projected/20de3f43-1fd9-47e0-9fae-d84370233104-kube-api-access-gbkn4\") pod \"observability-operator-cc5f78dfc-5jcdd\" (UID: \"20de3f43-1fd9-47e0-9fae-d84370233104\") " pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.159846 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/20de3f43-1fd9-47e0-9fae-d84370233104-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-5jcdd\" (UID: \"20de3f43-1fd9-47e0-9fae-d84370233104\") " pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.161257 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0568be7-153e-4be2-adb6-605c17922339-config-volume" (OuterVolumeSpecName: "config-volume") pod "b0568be7-153e-4be2-adb6-605c17922339" (UID: "b0568be7-153e-4be2-adb6-605c17922339"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.165471 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/20de3f43-1fd9-47e0-9fae-d84370233104-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-5jcdd\" (UID: \"20de3f43-1fd9-47e0-9fae-d84370233104\") " pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.165615 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0568be7-153e-4be2-adb6-605c17922339-kube-api-access-96j75" (OuterVolumeSpecName: "kube-api-access-96j75") pod "b0568be7-153e-4be2-adb6-605c17922339" (UID: "b0568be7-153e-4be2-adb6-605c17922339"). InnerVolumeSpecName "kube-api-access-96j75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.167048 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0568be7-153e-4be2-adb6-605c17922339-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b0568be7-153e-4be2-adb6-605c17922339" (UID: "b0568be7-153e-4be2-adb6-605c17922339"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.178027 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbkn4\" (UniqueName: \"kubernetes.io/projected/20de3f43-1fd9-47e0-9fae-d84370233104-kube-api-access-gbkn4\") pod \"observability-operator-cc5f78dfc-5jcdd\" (UID: \"20de3f43-1fd9-47e0-9fae-d84370233104\") " pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.225084 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.237142 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5"] Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.262333 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d7zm\" (UniqueName: \"kubernetes.io/projected/7d54e231-d091-4532-b5e9-5c2f27045e7e-kube-api-access-7d7zm\") pod \"perses-operator-54bc95c9fb-q42t7\" (UID: \"7d54e231-d091-4532-b5e9-5c2f27045e7e\") " pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.262407 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d54e231-d091-4532-b5e9-5c2f27045e7e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-q42t7\" (UID: \"7d54e231-d091-4532-b5e9-5c2f27045e7e\") " pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.262461 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0568be7-153e-4be2-adb6-605c17922339-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.262478 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0568be7-153e-4be2-adb6-605c17922339-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.262491 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96j75\" (UniqueName: \"kubernetes.io/projected/b0568be7-153e-4be2-adb6-605c17922339-kube-api-access-96j75\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.365837 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d54e231-d091-4532-b5e9-5c2f27045e7e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-q42t7\" (UID: \"7d54e231-d091-4532-b5e9-5c2f27045e7e\") " pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.365965 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d7zm\" (UniqueName: \"kubernetes.io/projected/7d54e231-d091-4532-b5e9-5c2f27045e7e-kube-api-access-7d7zm\") pod \"perses-operator-54bc95c9fb-q42t7\" (UID: \"7d54e231-d091-4532-b5e9-5c2f27045e7e\") " pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.367618 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d54e231-d091-4532-b5e9-5c2f27045e7e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-q42t7\" (UID: \"7d54e231-d091-4532-b5e9-5c2f27045e7e\") " pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.395927 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d7zm\" (UniqueName: \"kubernetes.io/projected/7d54e231-d091-4532-b5e9-5c2f27045e7e-kube-api-access-7d7zm\") pod \"perses-operator-54bc95c9fb-q42t7\" (UID: \"7d54e231-d091-4532-b5e9-5c2f27045e7e\") " pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.424275 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.479587 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn"] Oct 02 13:00:03 crc kubenswrapper[4710]: W1002 13:00:03.497989 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3df7c825_06b5_480f_920e_47216a812bfd.slice/crio-0ed6312a16512188ba48a1f19cdf3329173895f2e80a587568f17849c6c04de8 WatchSource:0}: Error finding container 0ed6312a16512188ba48a1f19cdf3329173895f2e80a587568f17849c6c04de8: Status 404 returned error can't find the container with id 0ed6312a16512188ba48a1f19cdf3329173895f2e80a587568f17849c6c04de8 Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.720255 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" event={"ID":"3df7c825-06b5-480f-920e-47216a812bfd","Type":"ContainerStarted","Data":"0ed6312a16512188ba48a1f19cdf3329173895f2e80a587568f17849c6c04de8"} Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.721867 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" event={"ID":"b0568be7-153e-4be2-adb6-605c17922339","Type":"ContainerDied","Data":"f65f93b27bffcf4b41a048899673d15a7009fc3088c566c41e932ca46c939d09"} Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.721899 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f65f93b27bffcf4b41a048899673d15a7009fc3088c566c41e932ca46c939d09" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.721959 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk" Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.732102 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" event={"ID":"4645af58-f9ab-4b88-a1fc-5a1ef95e330d","Type":"ContainerStarted","Data":"558e56561a9cb0ef2f006fc991744d557ed829452093aca3de3aef74fc4c5058"} Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.743009 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2"] Oct 02 13:00:03 crc kubenswrapper[4710]: W1002 13:00:03.749826 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd224b622_e6dd_47ce_868e_dfd2fe59d7c8.slice/crio-7806a5ff716de90f8fc994dde9aeb80b1f0f475371d763409f5f99c0789d9461 WatchSource:0}: Error finding container 7806a5ff716de90f8fc994dde9aeb80b1f0f475371d763409f5f99c0789d9461: Status 404 returned error can't find the container with id 7806a5ff716de90f8fc994dde9aeb80b1f0f475371d763409f5f99c0789d9461 Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.883321 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-5jcdd"] Oct 02 13:00:03 crc kubenswrapper[4710]: W1002 13:00:03.886634 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20de3f43_1fd9_47e0_9fae_d84370233104.slice/crio-198b3558554459e9790bcc475d38dc2a41e5167c3c394660fa3837220efa4a09 WatchSource:0}: Error finding container 198b3558554459e9790bcc475d38dc2a41e5167c3c394660fa3837220efa4a09: Status 404 returned error can't find the container with id 198b3558554459e9790bcc475d38dc2a41e5167c3c394660fa3837220efa4a09 Oct 02 13:00:03 crc kubenswrapper[4710]: I1002 13:00:03.933507 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-q42t7"] Oct 02 13:00:03 crc kubenswrapper[4710]: W1002 13:00:03.944337 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d54e231_d091_4532_b5e9_5c2f27045e7e.slice/crio-67a2bf41144564aeafa0c5f577c3eeb0790fe5cf1f6af55a38f381827244870a WatchSource:0}: Error finding container 67a2bf41144564aeafa0c5f577c3eeb0790fe5cf1f6af55a38f381827244870a: Status 404 returned error can't find the container with id 67a2bf41144564aeafa0c5f577c3eeb0790fe5cf1f6af55a38f381827244870a Oct 02 13:00:04 crc kubenswrapper[4710]: E1002 13:00:04.688502 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 13:00:04 crc kubenswrapper[4710]: I1002 13:00:04.737993 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" event={"ID":"7d54e231-d091-4532-b5e9-5c2f27045e7e","Type":"ContainerStarted","Data":"67a2bf41144564aeafa0c5f577c3eeb0790fe5cf1f6af55a38f381827244870a"} Oct 02 13:00:04 crc kubenswrapper[4710]: I1002 13:00:04.739317 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" event={"ID":"d224b622-e6dd-47ce-868e-dfd2fe59d7c8","Type":"ContainerStarted","Data":"7806a5ff716de90f8fc994dde9aeb80b1f0f475371d763409f5f99c0789d9461"} Oct 02 13:00:04 crc kubenswrapper[4710]: I1002 13:00:04.740286 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" event={"ID":"20de3f43-1fd9-47e0-9fae-d84370233104","Type":"ContainerStarted","Data":"198b3558554459e9790bcc475d38dc2a41e5167c3c394660fa3837220efa4a09"} Oct 02 13:00:14 crc kubenswrapper[4710]: E1002 13:00:14.804605 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e2e0c9_c0d7_40c0_92d5_1f7e00a3cc12.slice/crio-0bc5da267653d0cc69e9a4b137980a4f77064133c2485cbfee42d34d12b2aa3a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 13:00:14 crc kubenswrapper[4710]: I1002 13:00:14.841451 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" event={"ID":"4645af58-f9ab-4b88-a1fc-5a1ef95e330d","Type":"ContainerStarted","Data":"9de4153ed3c4821f5b27e527c4efdb437509e17a7468d945300d80b0bf7a203c"} Oct 02 13:00:14 crc kubenswrapper[4710]: I1002 13:00:14.857594 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" event={"ID":"3df7c825-06b5-480f-920e-47216a812bfd","Type":"ContainerStarted","Data":"f1dcbe49ca879da06cf74dbe084eaad73d5096957c180d9427cc4534c0afc7d7"} Oct 02 13:00:14 crc kubenswrapper[4710]: I1002 13:00:14.870268 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" event={"ID":"d224b622-e6dd-47ce-868e-dfd2fe59d7c8","Type":"ContainerStarted","Data":"09c9a211a9f57ca65a8fbd6c1e48505a66d278e3a045989d091984a19bdbf594"} Oct 02 13:00:14 crc kubenswrapper[4710]: I1002 13:00:14.886981 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-f8rs5" podStartSLOduration=2.364807491 podStartE2EDuration="12.886962902s" podCreationTimestamp="2025-10-02 13:00:02 +0000 UTC" firstStartedPulling="2025-10-02 13:00:03.268965903 +0000 UTC m=+647.375376786" lastFinishedPulling="2025-10-02 13:00:13.791121314 +0000 UTC m=+657.897532197" observedRunningTime="2025-10-02 13:00:14.865273783 +0000 UTC m=+658.971684676" watchObservedRunningTime="2025-10-02 13:00:14.886962902 +0000 UTC m=+658.993373785" Oct 02 13:00:14 crc kubenswrapper[4710]: I1002 13:00:14.888318 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn" podStartSLOduration=2.633590753 podStartE2EDuration="12.888313487s" podCreationTimestamp="2025-10-02 13:00:02 +0000 UTC" firstStartedPulling="2025-10-02 13:00:03.529912015 +0000 UTC m=+647.636322898" lastFinishedPulling="2025-10-02 13:00:13.784634749 +0000 UTC m=+657.891045632" observedRunningTime="2025-10-02 13:00:14.887942747 +0000 UTC m=+658.994353640" watchObservedRunningTime="2025-10-02 13:00:14.888313487 +0000 UTC m=+658.994724360" Oct 02 13:00:14 crc kubenswrapper[4710]: I1002 13:00:14.920989 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2" podStartSLOduration=2.865588891 podStartE2EDuration="12.920970224s" podCreationTimestamp="2025-10-02 13:00:02 +0000 UTC" firstStartedPulling="2025-10-02 13:00:03.753433409 +0000 UTC m=+647.859844292" lastFinishedPulling="2025-10-02 13:00:13.808814742 +0000 UTC m=+657.915225625" observedRunningTime="2025-10-02 13:00:14.91803957 +0000 UTC m=+659.024450463" watchObservedRunningTime="2025-10-02 13:00:14.920970224 +0000 UTC m=+659.027381107" Oct 02 13:00:15 crc kubenswrapper[4710]: I1002 13:00:15.878460 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" event={"ID":"7d54e231-d091-4532-b5e9-5c2f27045e7e","Type":"ContainerStarted","Data":"8ba4180670e0d2e15907f3d5ca5a18fa5b8801dabaec3d83671956e79f6cb79e"} Oct 02 13:00:15 crc kubenswrapper[4710]: I1002 13:00:15.893824 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" podStartSLOduration=1.428958177 podStartE2EDuration="12.893802086s" podCreationTimestamp="2025-10-02 13:00:03 +0000 UTC" firstStartedPulling="2025-10-02 13:00:03.946808719 +0000 UTC m=+648.053219602" lastFinishedPulling="2025-10-02 13:00:15.411652628 +0000 UTC m=+659.518063511" observedRunningTime="2025-10-02 13:00:15.893424496 +0000 UTC m=+659.999835379" watchObservedRunningTime="2025-10-02 13:00:15.893802086 +0000 UTC m=+660.000212969" Oct 02 13:00:16 crc kubenswrapper[4710]: I1002 13:00:16.884230 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:16 crc kubenswrapper[4710]: E1002 13:00:16.911161 4710 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/ddc0676de5843924e3f8d9857b0988f33c9d43c06503771c47a55756f876ddee/diff" to get inode usage: stat /var/lib/containers/storage/overlay/ddc0676de5843924e3f8d9857b0988f33c9d43c06503771c47a55756f876ddee/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/3.log" to get inode usage: stat /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8jhhq_93e2e0c9-c0d7-40c0-92d5-1f7e00a3cc12/ovnkube-controller/3.log: no such file or directory Oct 02 13:00:19 crc kubenswrapper[4710]: I1002 13:00:19.899962 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" event={"ID":"20de3f43-1fd9-47e0-9fae-d84370233104","Type":"ContainerStarted","Data":"7ec9d90fc0e65ea2a4f72a2a9b21332560a966943b991ed40603837c69778a4b"} Oct 02 13:00:19 crc kubenswrapper[4710]: I1002 13:00:19.900299 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:19 crc kubenswrapper[4710]: I1002 13:00:19.927424 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" Oct 02 13:00:19 crc kubenswrapper[4710]: I1002 13:00:19.937965 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-5jcdd" podStartSLOduration=2.988236389 podStartE2EDuration="17.937943974s" podCreationTimestamp="2025-10-02 13:00:02 +0000 UTC" firstStartedPulling="2025-10-02 13:00:03.889298322 +0000 UTC m=+647.995709205" lastFinishedPulling="2025-10-02 13:00:18.839005907 +0000 UTC m=+662.945416790" observedRunningTime="2025-10-02 13:00:19.93421848 +0000 UTC m=+664.040629373" watchObservedRunningTime="2025-10-02 13:00:19.937943974 +0000 UTC m=+664.044354857" Oct 02 13:00:23 crc kubenswrapper[4710]: I1002 13:00:23.427402 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-q42t7" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.324072 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh"] Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.329671 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.335947 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.351561 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh"] Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.454075 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.454142 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwfs\" (UniqueName: \"kubernetes.io/projected/4505d296-ddac-4be5-8ced-3dfc7e8eae59-kube-api-access-2lwfs\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.454212 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.555195 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.555272 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwfs\" (UniqueName: \"kubernetes.io/projected/4505d296-ddac-4be5-8ced-3dfc7e8eae59-kube-api-access-2lwfs\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.555320 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.555695 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.555813 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.576811 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwfs\" (UniqueName: \"kubernetes.io/projected/4505d296-ddac-4be5-8ced-3dfc7e8eae59-kube-api-access-2lwfs\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.664124 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:40 crc kubenswrapper[4710]: I1002 13:00:40.861183 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh"] Oct 02 13:00:41 crc kubenswrapper[4710]: I1002 13:00:41.013060 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" event={"ID":"4505d296-ddac-4be5-8ced-3dfc7e8eae59","Type":"ContainerStarted","Data":"953bbf416e364e1bd9a5c0262e4dc7a5a85d30a1c5d5a8289fae35f93415a513"} Oct 02 13:00:42 crc kubenswrapper[4710]: I1002 13:00:42.019120 4710 generic.go:334] "Generic (PLEG): container finished" podID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerID="63c43bc5366dd168e423f7bfde3954fb9bb2dffe2250fc0223cd12970d7fa5db" exitCode=0 Oct 02 13:00:42 crc kubenswrapper[4710]: I1002 13:00:42.019166 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" event={"ID":"4505d296-ddac-4be5-8ced-3dfc7e8eae59","Type":"ContainerDied","Data":"63c43bc5366dd168e423f7bfde3954fb9bb2dffe2250fc0223cd12970d7fa5db"} Oct 02 13:00:45 crc kubenswrapper[4710]: I1002 13:00:45.037879 4710 generic.go:334] "Generic (PLEG): container finished" podID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerID="1b372a6a45eb298ad834b9551866f38ff0a91a699e9d3184b3972633e68359dc" exitCode=0 Oct 02 13:00:45 crc kubenswrapper[4710]: I1002 13:00:45.037969 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" event={"ID":"4505d296-ddac-4be5-8ced-3dfc7e8eae59","Type":"ContainerDied","Data":"1b372a6a45eb298ad834b9551866f38ff0a91a699e9d3184b3972633e68359dc"} Oct 02 13:00:46 crc kubenswrapper[4710]: I1002 13:00:46.045467 4710 generic.go:334] "Generic (PLEG): container finished" podID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerID="45d7afeeb472c26f883a70a0b2d2b9e9045294219704df503c876d8bcc6339a6" exitCode=0 Oct 02 13:00:46 crc kubenswrapper[4710]: I1002 13:00:46.045527 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" event={"ID":"4505d296-ddac-4be5-8ced-3dfc7e8eae59","Type":"ContainerDied","Data":"45d7afeeb472c26f883a70a0b2d2b9e9045294219704df503c876d8bcc6339a6"} Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.304168 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.446413 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lwfs\" (UniqueName: \"kubernetes.io/projected/4505d296-ddac-4be5-8ced-3dfc7e8eae59-kube-api-access-2lwfs\") pod \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.446486 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-util\") pod \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.446513 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-bundle\") pod \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\" (UID: \"4505d296-ddac-4be5-8ced-3dfc7e8eae59\") " Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.447075 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-bundle" (OuterVolumeSpecName: "bundle") pod "4505d296-ddac-4be5-8ced-3dfc7e8eae59" (UID: "4505d296-ddac-4be5-8ced-3dfc7e8eae59"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.451886 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4505d296-ddac-4be5-8ced-3dfc7e8eae59-kube-api-access-2lwfs" (OuterVolumeSpecName: "kube-api-access-2lwfs") pod "4505d296-ddac-4be5-8ced-3dfc7e8eae59" (UID: "4505d296-ddac-4be5-8ced-3dfc7e8eae59"). InnerVolumeSpecName "kube-api-access-2lwfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.456362 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-util" (OuterVolumeSpecName: "util") pod "4505d296-ddac-4be5-8ced-3dfc7e8eae59" (UID: "4505d296-ddac-4be5-8ced-3dfc7e8eae59"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.548327 4710 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-util\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.548378 4710 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4505d296-ddac-4be5-8ced-3dfc7e8eae59-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:47 crc kubenswrapper[4710]: I1002 13:00:47.548395 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lwfs\" (UniqueName: \"kubernetes.io/projected/4505d296-ddac-4be5-8ced-3dfc7e8eae59-kube-api-access-2lwfs\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:48 crc kubenswrapper[4710]: I1002 13:00:48.058954 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" event={"ID":"4505d296-ddac-4be5-8ced-3dfc7e8eae59","Type":"ContainerDied","Data":"953bbf416e364e1bd9a5c0262e4dc7a5a85d30a1c5d5a8289fae35f93415a513"} Oct 02 13:00:48 crc kubenswrapper[4710]: I1002 13:00:48.059010 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="953bbf416e364e1bd9a5c0262e4dc7a5a85d30a1c5d5a8289fae35f93415a513" Oct 02 13:00:48 crc kubenswrapper[4710]: I1002 13:00:48.058987 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.322123 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf"] Oct 02 13:00:51 crc kubenswrapper[4710]: E1002 13:00:51.322948 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="extract" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.322962 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="extract" Oct 02 13:00:51 crc kubenswrapper[4710]: E1002 13:00:51.322971 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="util" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.322977 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="util" Oct 02 13:00:51 crc kubenswrapper[4710]: E1002 13:00:51.322987 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="pull" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.323006 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="pull" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.323161 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="4505d296-ddac-4be5-8ced-3dfc7e8eae59" containerName="extract" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.323697 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.326573 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.326979 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jvd89" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.328391 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.347271 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf"] Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.399504 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqnfm\" (UniqueName: \"kubernetes.io/projected/95074a87-510d-4d71-947e-b8cc858a29b2-kube-api-access-fqnfm\") pod \"nmstate-operator-858ddd8f98-9lmgf\" (UID: \"95074a87-510d-4d71-947e-b8cc858a29b2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.500925 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqnfm\" (UniqueName: \"kubernetes.io/projected/95074a87-510d-4d71-947e-b8cc858a29b2-kube-api-access-fqnfm\") pod \"nmstate-operator-858ddd8f98-9lmgf\" (UID: \"95074a87-510d-4d71-947e-b8cc858a29b2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.527921 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqnfm\" (UniqueName: \"kubernetes.io/projected/95074a87-510d-4d71-947e-b8cc858a29b2-kube-api-access-fqnfm\") pod \"nmstate-operator-858ddd8f98-9lmgf\" (UID: \"95074a87-510d-4d71-947e-b8cc858a29b2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.641662 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" Oct 02 13:00:51 crc kubenswrapper[4710]: I1002 13:00:51.937835 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf"] Oct 02 13:00:52 crc kubenswrapper[4710]: I1002 13:00:52.084435 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" event={"ID":"95074a87-510d-4d71-947e-b8cc858a29b2","Type":"ContainerStarted","Data":"14d6c5d2b105226a1918adc6b7333188a736c785f365f9587c2997922b19b07c"} Oct 02 13:00:53 crc kubenswrapper[4710]: I1002 13:00:53.753969 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gqwlv"] Oct 02 13:00:53 crc kubenswrapper[4710]: I1002 13:00:53.755427 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" podUID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" containerName="controller-manager" containerID="cri-o://e32ab99cfa7e73c0a4994afc8b467e0231928e090472e8c9b6c8e4d1062824e0" gracePeriod=30 Oct 02 13:00:53 crc kubenswrapper[4710]: I1002 13:00:53.860882 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6"] Oct 02 13:00:53 crc kubenswrapper[4710]: I1002 13:00:53.861123 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" podUID="f176176a-261c-4a3a-a88f-a8ae63405edf" containerName="route-controller-manager" containerID="cri-o://8e9f15d78a87dcd63369cb886a1ccb184b9b6a43a31bcd13e75a829184bb5a58" gracePeriod=30 Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.100770 4710 generic.go:334] "Generic (PLEG): container finished" podID="f176176a-261c-4a3a-a88f-a8ae63405edf" containerID="8e9f15d78a87dcd63369cb886a1ccb184b9b6a43a31bcd13e75a829184bb5a58" exitCode=0 Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.100835 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" event={"ID":"f176176a-261c-4a3a-a88f-a8ae63405edf","Type":"ContainerDied","Data":"8e9f15d78a87dcd63369cb886a1ccb184b9b6a43a31bcd13e75a829184bb5a58"} Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.102589 4710 generic.go:334] "Generic (PLEG): container finished" podID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" containerID="e32ab99cfa7e73c0a4994afc8b467e0231928e090472e8c9b6c8e4d1062824e0" exitCode=0 Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.102612 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" event={"ID":"6cdd48c4-7366-4e02-9cf9-78cb66e42249","Type":"ContainerDied","Data":"e32ab99cfa7e73c0a4994afc8b467e0231928e090472e8c9b6c8e4d1062824e0"} Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.271117 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.338592 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f176176a-261c-4a3a-a88f-a8ae63405edf-serving-cert\") pod \"f176176a-261c-4a3a-a88f-a8ae63405edf\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.338695 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-config\") pod \"f176176a-261c-4a3a-a88f-a8ae63405edf\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.338767 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-client-ca\") pod \"f176176a-261c-4a3a-a88f-a8ae63405edf\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.338806 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hww5p\" (UniqueName: \"kubernetes.io/projected/f176176a-261c-4a3a-a88f-a8ae63405edf-kube-api-access-hww5p\") pod \"f176176a-261c-4a3a-a88f-a8ae63405edf\" (UID: \"f176176a-261c-4a3a-a88f-a8ae63405edf\") " Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.340570 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-config" (OuterVolumeSpecName: "config") pod "f176176a-261c-4a3a-a88f-a8ae63405edf" (UID: "f176176a-261c-4a3a-a88f-a8ae63405edf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.340662 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-client-ca" (OuterVolumeSpecName: "client-ca") pod "f176176a-261c-4a3a-a88f-a8ae63405edf" (UID: "f176176a-261c-4a3a-a88f-a8ae63405edf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.344341 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f176176a-261c-4a3a-a88f-a8ae63405edf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f176176a-261c-4a3a-a88f-a8ae63405edf" (UID: "f176176a-261c-4a3a-a88f-a8ae63405edf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.344689 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f176176a-261c-4a3a-a88f-a8ae63405edf-kube-api-access-hww5p" (OuterVolumeSpecName: "kube-api-access-hww5p") pod "f176176a-261c-4a3a-a88f-a8ae63405edf" (UID: "f176176a-261c-4a3a-a88f-a8ae63405edf"). InnerVolumeSpecName "kube-api-access-hww5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.439797 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f176176a-261c-4a3a-a88f-a8ae63405edf-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.439837 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.439848 4710 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f176176a-261c-4a3a-a88f-a8ae63405edf-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.439859 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hww5p\" (UniqueName: \"kubernetes.io/projected/f176176a-261c-4a3a-a88f-a8ae63405edf-kube-api-access-hww5p\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:54 crc kubenswrapper[4710]: I1002 13:00:54.981251 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.047144 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8q9p\" (UniqueName: \"kubernetes.io/projected/6cdd48c4-7366-4e02-9cf9-78cb66e42249-kube-api-access-q8q9p\") pod \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.047190 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-config\") pod \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.047241 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdd48c4-7366-4e02-9cf9-78cb66e42249-serving-cert\") pod \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.047260 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-client-ca\") pod \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.047289 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-proxy-ca-bundles\") pod \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\" (UID: \"6cdd48c4-7366-4e02-9cf9-78cb66e42249\") " Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.048282 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-client-ca" (OuterVolumeSpecName: "client-ca") pod "6cdd48c4-7366-4e02-9cf9-78cb66e42249" (UID: "6cdd48c4-7366-4e02-9cf9-78cb66e42249"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.048620 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-config" (OuterVolumeSpecName: "config") pod "6cdd48c4-7366-4e02-9cf9-78cb66e42249" (UID: "6cdd48c4-7366-4e02-9cf9-78cb66e42249"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.048865 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6cdd48c4-7366-4e02-9cf9-78cb66e42249" (UID: "6cdd48c4-7366-4e02-9cf9-78cb66e42249"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.052636 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cdd48c4-7366-4e02-9cf9-78cb66e42249-kube-api-access-q8q9p" (OuterVolumeSpecName: "kube-api-access-q8q9p") pod "6cdd48c4-7366-4e02-9cf9-78cb66e42249" (UID: "6cdd48c4-7366-4e02-9cf9-78cb66e42249"). InnerVolumeSpecName "kube-api-access-q8q9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.052644 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cdd48c4-7366-4e02-9cf9-78cb66e42249-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6cdd48c4-7366-4e02-9cf9-78cb66e42249" (UID: "6cdd48c4-7366-4e02-9cf9-78cb66e42249"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.108569 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" event={"ID":"f176176a-261c-4a3a-a88f-a8ae63405edf","Type":"ContainerDied","Data":"ab2eaa0649551d0b24f625ec782731e826cb4f51a207d0552ba7a0e571505fb2"} Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.108627 4710 scope.go:117] "RemoveContainer" containerID="8e9f15d78a87dcd63369cb886a1ccb184b9b6a43a31bcd13e75a829184bb5a58" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.108625 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.110335 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" event={"ID":"6cdd48c4-7366-4e02-9cf9-78cb66e42249","Type":"ContainerDied","Data":"a13c1c9604f1f15c6aaf4ff7845cc434030bef0f510b96ae69dcd02c97164616"} Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.110422 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gqwlv" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.125172 4710 scope.go:117] "RemoveContainer" containerID="e32ab99cfa7e73c0a4994afc8b467e0231928e090472e8c9b6c8e4d1062824e0" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.125531 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.128579 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftnw6"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.146659 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gqwlv"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.149426 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8q9p\" (UniqueName: \"kubernetes.io/projected/6cdd48c4-7366-4e02-9cf9-78cb66e42249-kube-api-access-q8q9p\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.149611 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.149683 4710 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdd48c4-7366-4e02-9cf9-78cb66e42249-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.149768 4710 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.149906 4710 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cdd48c4-7366-4e02-9cf9-78cb66e42249-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.150707 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gqwlv"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.597067 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx"] Oct 02 13:00:55 crc kubenswrapper[4710]: E1002 13:00:55.597312 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f176176a-261c-4a3a-a88f-a8ae63405edf" containerName="route-controller-manager" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.597330 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="f176176a-261c-4a3a-a88f-a8ae63405edf" containerName="route-controller-manager" Oct 02 13:00:55 crc kubenswrapper[4710]: E1002 13:00:55.597347 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" containerName="controller-manager" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.597355 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" containerName="controller-manager" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.597488 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="f176176a-261c-4a3a-a88f-a8ae63405edf" containerName="route-controller-manager" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.597502 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" containerName="controller-manager" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.597974 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.602379 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f886d8885-wvbvb"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.603410 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.605668 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.606723 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.606834 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.606925 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.606988 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.607107 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.607229 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.607103 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.607108 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.609542 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.609569 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.613674 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.615633 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f886d8885-wvbvb"] Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.617014 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.617053 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.655847 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8b4p\" (UniqueName: \"kubernetes.io/projected/4923f3a5-1e2e-46a7-b97b-009d529b038e-kube-api-access-r8b4p\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.655906 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-config\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.655946 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-client-ca\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.655976 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4923f3a5-1e2e-46a7-b97b-009d529b038e-config\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.656025 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-serving-cert\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.656109 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjkvg\" (UniqueName: \"kubernetes.io/projected/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-kube-api-access-rjkvg\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.656135 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4923f3a5-1e2e-46a7-b97b-009d529b038e-serving-cert\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.656170 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4923f3a5-1e2e-46a7-b97b-009d529b038e-client-ca\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.656194 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-proxy-ca-bundles\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.757484 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-serving-cert\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.757725 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjkvg\" (UniqueName: \"kubernetes.io/projected/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-kube-api-access-rjkvg\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.757870 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4923f3a5-1e2e-46a7-b97b-009d529b038e-serving-cert\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.757966 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4923f3a5-1e2e-46a7-b97b-009d529b038e-client-ca\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.758039 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-proxy-ca-bundles\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.758119 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8b4p\" (UniqueName: \"kubernetes.io/projected/4923f3a5-1e2e-46a7-b97b-009d529b038e-kube-api-access-r8b4p\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.758203 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-config\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.758281 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-client-ca\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.758351 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4923f3a5-1e2e-46a7-b97b-009d529b038e-config\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.758949 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4923f3a5-1e2e-46a7-b97b-009d529b038e-client-ca\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.759004 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-client-ca\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.759047 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-proxy-ca-bundles\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.759534 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4923f3a5-1e2e-46a7-b97b-009d529b038e-config\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.760276 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-config\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.762204 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4923f3a5-1e2e-46a7-b97b-009d529b038e-serving-cert\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.762244 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-serving-cert\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.777407 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8b4p\" (UniqueName: \"kubernetes.io/projected/4923f3a5-1e2e-46a7-b97b-009d529b038e-kube-api-access-r8b4p\") pod \"route-controller-manager-578bbcf5c5-dz4jx\" (UID: \"4923f3a5-1e2e-46a7-b97b-009d529b038e\") " pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.777638 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjkvg\" (UniqueName: \"kubernetes.io/projected/6c6dd9a9-232c-42f4-bcfb-a203e0da62f6-kube-api-access-rjkvg\") pod \"controller-manager-f886d8885-wvbvb\" (UID: \"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6\") " pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.914653 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:55 crc kubenswrapper[4710]: I1002 13:00:55.921518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:56 crc kubenswrapper[4710]: I1002 13:00:56.120083 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" event={"ID":"95074a87-510d-4d71-947e-b8cc858a29b2","Type":"ContainerStarted","Data":"efdd61031ccaf8c969b8618797e2351515dcb1d1324a6eb32e0f24446a85b0de"} Oct 02 13:00:56 crc kubenswrapper[4710]: I1002 13:00:56.155127 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-9lmgf" podStartSLOduration=2.131131746 podStartE2EDuration="5.155104434s" podCreationTimestamp="2025-10-02 13:00:51 +0000 UTC" firstStartedPulling="2025-10-02 13:00:51.966096594 +0000 UTC m=+696.072507477" lastFinishedPulling="2025-10-02 13:00:54.990069282 +0000 UTC m=+699.096480165" observedRunningTime="2025-10-02 13:00:56.148344722 +0000 UTC m=+700.254755605" watchObservedRunningTime="2025-10-02 13:00:56.155104434 +0000 UTC m=+700.261515317" Oct 02 13:00:56 crc kubenswrapper[4710]: I1002 13:00:56.282629 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f886d8885-wvbvb"] Oct 02 13:00:56 crc kubenswrapper[4710]: I1002 13:00:56.360851 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx"] Oct 02 13:00:56 crc kubenswrapper[4710]: W1002 13:00:56.362627 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4923f3a5_1e2e_46a7_b97b_009d529b038e.slice/crio-5772081b8816f9b1bdbcb0844ac90078e8eb877cd1457a26e6d959b6f1dd4e2c WatchSource:0}: Error finding container 5772081b8816f9b1bdbcb0844ac90078e8eb877cd1457a26e6d959b6f1dd4e2c: Status 404 returned error can't find the container with id 5772081b8816f9b1bdbcb0844ac90078e8eb877cd1457a26e6d959b6f1dd4e2c Oct 02 13:00:56 crc kubenswrapper[4710]: I1002 13:00:56.911490 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cdd48c4-7366-4e02-9cf9-78cb66e42249" path="/var/lib/kubelet/pods/6cdd48c4-7366-4e02-9cf9-78cb66e42249/volumes" Oct 02 13:00:56 crc kubenswrapper[4710]: I1002 13:00:56.912552 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f176176a-261c-4a3a-a88f-a8ae63405edf" path="/var/lib/kubelet/pods/f176176a-261c-4a3a-a88f-a8ae63405edf/volumes" Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.130208 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" event={"ID":"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6","Type":"ContainerStarted","Data":"9761230d07885734ebdf132c230d5af3899cc4dc634fbed6914def74ff57da91"} Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.130299 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.130338 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" event={"ID":"6c6dd9a9-232c-42f4-bcfb-a203e0da62f6","Type":"ContainerStarted","Data":"52817ccbd70e323a8236fbfd52d81298f7fc3b04d55b07a7952ff9d8f790c88e"} Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.131973 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" event={"ID":"4923f3a5-1e2e-46a7-b97b-009d529b038e","Type":"ContainerStarted","Data":"8fb5fb3d8b1eda115b199b31221196d72f0cb3f0af0ff5ac9be65b14105541c9"} Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.132046 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" event={"ID":"4923f3a5-1e2e-46a7-b97b-009d529b038e","Type":"ContainerStarted","Data":"5772081b8816f9b1bdbcb0844ac90078e8eb877cd1457a26e6d959b6f1dd4e2c"} Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.132103 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.135454 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.136855 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.150534 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f886d8885-wvbvb" podStartSLOduration=4.150511547 podStartE2EDuration="4.150511547s" podCreationTimestamp="2025-10-02 13:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:00:57.146193397 +0000 UTC m=+701.252604290" watchObservedRunningTime="2025-10-02 13:00:57.150511547 +0000 UTC m=+701.256922440" Oct 02 13:00:57 crc kubenswrapper[4710]: I1002 13:00:57.194616 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-578bbcf5c5-dz4jx" podStartSLOduration=4.194595954 podStartE2EDuration="4.194595954s" podCreationTimestamp="2025-10-02 13:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:00:57.189989607 +0000 UTC m=+701.296400500" watchObservedRunningTime="2025-10-02 13:00:57.194595954 +0000 UTC m=+701.301006837" Oct 02 13:00:59 crc kubenswrapper[4710]: I1002 13:00:59.630070 4710 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.246768 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.248168 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.249879 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-tmk78" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.259658 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.269149 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.270371 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.276256 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.294405 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-xzv7r"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.295260 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.303119 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334087 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt7rb\" (UniqueName: \"kubernetes.io/projected/f680c7e5-dff2-426c-9edd-387bfc613153-kube-api-access-zt7rb\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334165 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-nmstate-lock\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334186 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-ovs-socket\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334293 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f680c7e5-dff2-426c-9edd-387bfc613153-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334327 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-dbus-socket\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334342 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmjfn\" (UniqueName: \"kubernetes.io/projected/fdd7df24-9aea-4c00-9b0a-424cb370d89f-kube-api-access-wmjfn\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.334369 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79sxg\" (UniqueName: \"kubernetes.io/projected/34ee392d-77a8-44d3-9bde-ccbb875628fc-kube-api-access-79sxg\") pod \"nmstate-metrics-fdff9cb8d-f5lpb\" (UID: \"34ee392d-77a8-44d3-9bde-ccbb875628fc\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.425602 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.426282 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.427892 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.428093 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9mng8" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.428241 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435734 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-dbus-socket\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435786 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmjfn\" (UniqueName: \"kubernetes.io/projected/fdd7df24-9aea-4c00-9b0a-424cb370d89f-kube-api-access-wmjfn\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435818 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79sxg\" (UniqueName: \"kubernetes.io/projected/34ee392d-77a8-44d3-9bde-ccbb875628fc-kube-api-access-79sxg\") pod \"nmstate-metrics-fdff9cb8d-f5lpb\" (UID: \"34ee392d-77a8-44d3-9bde-ccbb875628fc\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435852 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt7rb\" (UniqueName: \"kubernetes.io/projected/f680c7e5-dff2-426c-9edd-387bfc613153-kube-api-access-zt7rb\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435869 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-nmstate-lock\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435884 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-ovs-socket\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.435922 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f680c7e5-dff2-426c-9edd-387bfc613153-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: E1002 13:01:01.436032 4710 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 13:01:01 crc kubenswrapper[4710]: E1002 13:01:01.436076 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f680c7e5-dff2-426c-9edd-387bfc613153-tls-key-pair podName:f680c7e5-dff2-426c-9edd-387bfc613153 nodeName:}" failed. No retries permitted until 2025-10-02 13:01:01.936059552 +0000 UTC m=+706.042470435 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f680c7e5-dff2-426c-9edd-387bfc613153-tls-key-pair") pod "nmstate-webhook-6cdbc54649-dtspq" (UID: "f680c7e5-dff2-426c-9edd-387bfc613153") : secret "openshift-nmstate-webhook" not found Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.436447 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-dbus-socket\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.436832 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-nmstate-lock\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.436865 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fdd7df24-9aea-4c00-9b0a-424cb370d89f-ovs-socket\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.462885 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79sxg\" (UniqueName: \"kubernetes.io/projected/34ee392d-77a8-44d3-9bde-ccbb875628fc-kube-api-access-79sxg\") pod \"nmstate-metrics-fdff9cb8d-f5lpb\" (UID: \"34ee392d-77a8-44d3-9bde-ccbb875628fc\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.463482 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt7rb\" (UniqueName: \"kubernetes.io/projected/f680c7e5-dff2-426c-9edd-387bfc613153-kube-api-access-zt7rb\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.465493 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmjfn\" (UniqueName: \"kubernetes.io/projected/fdd7df24-9aea-4c00-9b0a-424cb370d89f-kube-api-access-wmjfn\") pod \"nmstate-handler-xzv7r\" (UID: \"fdd7df24-9aea-4c00-9b0a-424cb370d89f\") " pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.475298 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.537372 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fc6708be-e01d-45d3-84b9-34f09476898f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.537443 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6708be-e01d-45d3-84b9-34f09476898f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.537501 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqq4x\" (UniqueName: \"kubernetes.io/projected/fc6708be-e01d-45d3-84b9-34f09476898f-kube-api-access-lqq4x\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.572487 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.613210 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.638259 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fc6708be-e01d-45d3-84b9-34f09476898f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.638318 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6708be-e01d-45d3-84b9-34f09476898f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.638370 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqq4x\" (UniqueName: \"kubernetes.io/projected/fc6708be-e01d-45d3-84b9-34f09476898f-kube-api-access-lqq4x\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: E1002 13:01:01.638532 4710 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 02 13:01:01 crc kubenswrapper[4710]: E1002 13:01:01.638623 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc6708be-e01d-45d3-84b9-34f09476898f-plugin-serving-cert podName:fc6708be-e01d-45d3-84b9-34f09476898f nodeName:}" failed. No retries permitted until 2025-10-02 13:01:02.138595944 +0000 UTC m=+706.245006827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/fc6708be-e01d-45d3-84b9-34f09476898f-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-2m97m" (UID: "fc6708be-e01d-45d3-84b9-34f09476898f") : secret "plugin-serving-cert" not found Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.639963 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fc6708be-e01d-45d3-84b9-34f09476898f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.640209 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5d696f7c8c-j4f8c"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.641432 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.666296 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d696f7c8c-j4f8c"] Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.674654 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqq4x\" (UniqueName: \"kubernetes.io/projected/fc6708be-e01d-45d3-84b9-34f09476898f-kube-api-access-lqq4x\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.739817 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbf5n\" (UniqueName: \"kubernetes.io/projected/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-kube-api-access-bbf5n\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.739870 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-trusted-ca-bundle\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.739925 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-oauth-config\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.740001 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-serving-cert\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.740024 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-oauth-serving-cert\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.740052 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-config\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.740100 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-service-ca\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.840864 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbf5n\" (UniqueName: \"kubernetes.io/projected/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-kube-api-access-bbf5n\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.840923 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-trusted-ca-bundle\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.840967 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-oauth-config\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.841027 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-serving-cert\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.841049 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-oauth-serving-cert\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.841090 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-config\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.841118 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-service-ca\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.842523 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-service-ca\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.843893 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-config\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.844451 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-oauth-serving-cert\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.845225 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-trusted-ca-bundle\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.847272 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-serving-cert\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.849495 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-console-oauth-config\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.861810 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbf5n\" (UniqueName: \"kubernetes.io/projected/55fa34b6-ab3e-4866-bac3-2cfb9e84a98e-kube-api-access-bbf5n\") pod \"console-5d696f7c8c-j4f8c\" (UID: \"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e\") " pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.942653 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f680c7e5-dff2-426c-9edd-387bfc613153-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.948463 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f680c7e5-dff2-426c-9edd-387bfc613153-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dtspq\" (UID: \"f680c7e5-dff2-426c-9edd-387bfc613153\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:01 crc kubenswrapper[4710]: I1002 13:01:01.968352 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.070581 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb"] Oct 02 13:01:02 crc kubenswrapper[4710]: W1002 13:01:02.079459 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34ee392d_77a8_44d3_9bde_ccbb875628fc.slice/crio-f9c45553be0b654369e5369bcb53b6dc641be251c47729683f8526e66f93a43d WatchSource:0}: Error finding container f9c45553be0b654369e5369bcb53b6dc641be251c47729683f8526e66f93a43d: Status 404 returned error can't find the container with id f9c45553be0b654369e5369bcb53b6dc641be251c47729683f8526e66f93a43d Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.145687 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6708be-e01d-45d3-84b9-34f09476898f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.150076 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6708be-e01d-45d3-84b9-34f09476898f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2m97m\" (UID: \"fc6708be-e01d-45d3-84b9-34f09476898f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.159225 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xzv7r" event={"ID":"fdd7df24-9aea-4c00-9b0a-424cb370d89f","Type":"ContainerStarted","Data":"eb0e4c5a685e7e07de703b2bea9e0ecfec5575315ad84c8ca3b26d6aa0f485e5"} Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.160474 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" event={"ID":"34ee392d-77a8-44d3-9bde-ccbb875628fc","Type":"ContainerStarted","Data":"f9c45553be0b654369e5369bcb53b6dc641be251c47729683f8526e66f93a43d"} Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.192640 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.343735 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.404965 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d696f7c8c-j4f8c"] Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.645114 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq"] Oct 02 13:01:02 crc kubenswrapper[4710]: I1002 13:01:02.861838 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m"] Oct 02 13:01:03 crc kubenswrapper[4710]: I1002 13:01:03.170825 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d696f7c8c-j4f8c" event={"ID":"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e","Type":"ContainerStarted","Data":"fc0a3b5c5f2927d361cf1cbdad9be2bddaa544772b6df4c3fb892fd298caac0e"} Oct 02 13:01:03 crc kubenswrapper[4710]: I1002 13:01:03.170886 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d696f7c8c-j4f8c" event={"ID":"55fa34b6-ab3e-4866-bac3-2cfb9e84a98e","Type":"ContainerStarted","Data":"4d8796679d54f1171d9e7582b78da61ecb5e439baf8065019f01726b8b0461cf"} Oct 02 13:01:03 crc kubenswrapper[4710]: I1002 13:01:03.172218 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" event={"ID":"f680c7e5-dff2-426c-9edd-387bfc613153","Type":"ContainerStarted","Data":"cd04005e7cecb1a3e71146aaaf2a01d38051ca51007210efadeccfac9697df48"} Oct 02 13:01:03 crc kubenswrapper[4710]: I1002 13:01:03.173221 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" event={"ID":"fc6708be-e01d-45d3-84b9-34f09476898f","Type":"ContainerStarted","Data":"14c276b12e39c0b246f762bd1a7dc387cba3127769bd1d6c2e8e42c8ea9578e1"} Oct 02 13:01:03 crc kubenswrapper[4710]: I1002 13:01:03.196876 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5d696f7c8c-j4f8c" podStartSLOduration=2.1968490210000002 podStartE2EDuration="2.196849021s" podCreationTimestamp="2025-10-02 13:01:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:01:03.189769961 +0000 UTC m=+707.296180854" watchObservedRunningTime="2025-10-02 13:01:03.196849021 +0000 UTC m=+707.303259904" Oct 02 13:01:06 crc kubenswrapper[4710]: I1002 13:01:06.195272 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" event={"ID":"34ee392d-77a8-44d3-9bde-ccbb875628fc","Type":"ContainerStarted","Data":"bc51d129ee78d61dcede1e93a8b03aff1dcb3c3ea6deab09e87f52c8f3454e37"} Oct 02 13:01:06 crc kubenswrapper[4710]: I1002 13:01:06.198616 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xzv7r" event={"ID":"fdd7df24-9aea-4c00-9b0a-424cb370d89f","Type":"ContainerStarted","Data":"0fbec99a9579cb27b7858f86ed40ec56aa7ab599b9d929bb6df02ac9f3910d71"} Oct 02 13:01:06 crc kubenswrapper[4710]: I1002 13:01:06.199195 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:06 crc kubenswrapper[4710]: I1002 13:01:06.218999 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-xzv7r" podStartSLOduration=1.03661661 podStartE2EDuration="5.218982541s" podCreationTimestamp="2025-10-02 13:01:01 +0000 UTC" firstStartedPulling="2025-10-02 13:01:01.674398232 +0000 UTC m=+705.780809115" lastFinishedPulling="2025-10-02 13:01:05.856764163 +0000 UTC m=+709.963175046" observedRunningTime="2025-10-02 13:01:06.214738174 +0000 UTC m=+710.321149057" watchObservedRunningTime="2025-10-02 13:01:06.218982541 +0000 UTC m=+710.325393424" Oct 02 13:01:07 crc kubenswrapper[4710]: I1002 13:01:07.207819 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" event={"ID":"fc6708be-e01d-45d3-84b9-34f09476898f","Type":"ContainerStarted","Data":"87aa60c364f2594186850448391d3a9c9272fa37d7eabe327622df979a2870f7"} Oct 02 13:01:07 crc kubenswrapper[4710]: I1002 13:01:07.210198 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" event={"ID":"f680c7e5-dff2-426c-9edd-387bfc613153","Type":"ContainerStarted","Data":"32c1f059fd9a3abdc76c4ec7d14d8c6992633bab2ffdafe71a536ef5deadd78a"} Oct 02 13:01:07 crc kubenswrapper[4710]: I1002 13:01:07.230024 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2m97m" podStartSLOduration=2.1014438970000002 podStartE2EDuration="6.229996725s" podCreationTimestamp="2025-10-02 13:01:01 +0000 UTC" firstStartedPulling="2025-10-02 13:01:02.873035335 +0000 UTC m=+706.979446218" lastFinishedPulling="2025-10-02 13:01:07.001588163 +0000 UTC m=+711.107999046" observedRunningTime="2025-10-02 13:01:07.227810634 +0000 UTC m=+711.334221527" watchObservedRunningTime="2025-10-02 13:01:07.229996725 +0000 UTC m=+711.336407618" Oct 02 13:01:07 crc kubenswrapper[4710]: I1002 13:01:07.249426 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" podStartSLOduration=2.81716188 podStartE2EDuration="6.249408643s" podCreationTimestamp="2025-10-02 13:01:01 +0000 UTC" firstStartedPulling="2025-10-02 13:01:02.659101914 +0000 UTC m=+706.765512787" lastFinishedPulling="2025-10-02 13:01:06.091348667 +0000 UTC m=+710.197759550" observedRunningTime="2025-10-02 13:01:07.248305897 +0000 UTC m=+711.354716790" watchObservedRunningTime="2025-10-02 13:01:07.249408643 +0000 UTC m=+711.355819526" Oct 02 13:01:08 crc kubenswrapper[4710]: I1002 13:01:08.218502 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:09 crc kubenswrapper[4710]: I1002 13:01:09.227307 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" event={"ID":"34ee392d-77a8-44d3-9bde-ccbb875628fc","Type":"ContainerStarted","Data":"dbbfa99c935c7347bfebb3e9f5c5afcddedd41e49c8388f132624595025f0fa0"} Oct 02 13:01:09 crc kubenswrapper[4710]: I1002 13:01:09.250500 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f5lpb" podStartSLOduration=1.48146719 podStartE2EDuration="8.250477622s" podCreationTimestamp="2025-10-02 13:01:01 +0000 UTC" firstStartedPulling="2025-10-02 13:01:02.083891288 +0000 UTC m=+706.190302171" lastFinishedPulling="2025-10-02 13:01:08.85290172 +0000 UTC m=+712.959312603" observedRunningTime="2025-10-02 13:01:09.246652462 +0000 UTC m=+713.353063365" watchObservedRunningTime="2025-10-02 13:01:09.250477622 +0000 UTC m=+713.356888515" Oct 02 13:01:11 crc kubenswrapper[4710]: I1002 13:01:11.642156 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-xzv7r" Oct 02 13:01:11 crc kubenswrapper[4710]: I1002 13:01:11.968567 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:11 crc kubenswrapper[4710]: I1002 13:01:11.969052 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:11 crc kubenswrapper[4710]: I1002 13:01:11.978862 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:12 crc kubenswrapper[4710]: I1002 13:01:12.249463 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5d696f7c8c-j4f8c" Oct 02 13:01:12 crc kubenswrapper[4710]: I1002 13:01:12.302950 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mrtmb"] Oct 02 13:01:22 crc kubenswrapper[4710]: I1002 13:01:22.198479 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dtspq" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.341120 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mrtmb" podUID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" containerName="console" containerID="cri-o://dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3" gracePeriod=15 Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.376465 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8"] Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.377703 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.379610 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.388057 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8"] Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.569900 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scm6p\" (UniqueName: \"kubernetes.io/projected/1675361e-20c2-46ab-8d3a-0b28c519fc3d-kube-api-access-scm6p\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.569988 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.570016 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.671396 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scm6p\" (UniqueName: \"kubernetes.io/projected/1675361e-20c2-46ab-8d3a-0b28c519fc3d-kube-api-access-scm6p\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.671461 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.671487 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.672104 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.672241 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.695206 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scm6p\" (UniqueName: \"kubernetes.io/projected/1675361e-20c2-46ab-8d3a-0b28c519fc3d-kube-api-access-scm6p\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.740300 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mrtmb_1ef751b2-7d8e-4caa-ad61-d8544a501c6b/console/0.log" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.740376 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.742314 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873142 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-service-ca\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873188 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-trusted-ca-bundle\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873298 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-serving-cert\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873334 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-config\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873357 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-oauth-serving-cert\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873389 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-oauth-config\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.873408 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g65nh\" (UniqueName: \"kubernetes.io/projected/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-kube-api-access-g65nh\") pod \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\" (UID: \"1ef751b2-7d8e-4caa-ad61-d8544a501c6b\") " Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.874177 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.874265 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-service-ca" (OuterVolumeSpecName: "service-ca") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.874382 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-config" (OuterVolumeSpecName: "console-config") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.874884 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.880466 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-kube-api-access-g65nh" (OuterVolumeSpecName: "kube-api-access-g65nh") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "kube-api-access-g65nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.881664 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.881963 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "1ef751b2-7d8e-4caa-ad61-d8544a501c6b" (UID: "1ef751b2-7d8e-4caa-ad61-d8544a501c6b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974687 4710 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974723 4710 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974733 4710 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974741 4710 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974762 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g65nh\" (UniqueName: \"kubernetes.io/projected/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-kube-api-access-g65nh\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974773 4710 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:37 crc kubenswrapper[4710]: I1002 13:01:37.974780 4710 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef751b2-7d8e-4caa-ad61-d8544a501c6b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.166880 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8"] Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.412219 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mrtmb_1ef751b2-7d8e-4caa-ad61-d8544a501c6b/console/0.log" Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.412275 4710 generic.go:334] "Generic (PLEG): container finished" podID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" containerID="dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3" exitCode=2 Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.412337 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mrtmb" event={"ID":"1ef751b2-7d8e-4caa-ad61-d8544a501c6b","Type":"ContainerDied","Data":"dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3"} Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.412367 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mrtmb" event={"ID":"1ef751b2-7d8e-4caa-ad61-d8544a501c6b","Type":"ContainerDied","Data":"2035507cfec1479c3405989148a50bafda2d592b692195b1f1a8163e53d7db71"} Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.412371 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mrtmb" Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.412388 4710 scope.go:117] "RemoveContainer" containerID="dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3" Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.414712 4710 generic.go:334] "Generic (PLEG): container finished" podID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerID="9d8f088e7541eff0ff3488e605bb3eb3bc3b4a3225cb792b67dff96e15bcbc3e" exitCode=0 Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.414824 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" event={"ID":"1675361e-20c2-46ab-8d3a-0b28c519fc3d","Type":"ContainerDied","Data":"9d8f088e7541eff0ff3488e605bb3eb3bc3b4a3225cb792b67dff96e15bcbc3e"} Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.414858 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" event={"ID":"1675361e-20c2-46ab-8d3a-0b28c519fc3d","Type":"ContainerStarted","Data":"cfa4281b82d48105cdef651ab273240f3ee57d28d9a86bac4bcd00f9776fa345"} Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.455081 4710 scope.go:117] "RemoveContainer" containerID="dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3" Oct 02 13:01:38 crc kubenswrapper[4710]: E1002 13:01:38.455840 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3\": container with ID starting with dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3 not found: ID does not exist" containerID="dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3" Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.455905 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3"} err="failed to get container status \"dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3\": rpc error: code = NotFound desc = could not find container \"dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3\": container with ID starting with dcdce4cecc9c7db9cfba9bbe5a7eb7ad796d86cb7c8d4ef0a57fcc58f40a1db3 not found: ID does not exist" Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.489911 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mrtmb"] Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.493630 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mrtmb"] Oct 02 13:01:38 crc kubenswrapper[4710]: I1002 13:01:38.917715 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" path="/var/lib/kubelet/pods/1ef751b2-7d8e-4caa-ad61-d8544a501c6b/volumes" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.748660 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fkb9d"] Oct 02 13:01:39 crc kubenswrapper[4710]: E1002 13:01:39.749514 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" containerName="console" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.749549 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" containerName="console" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.749883 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef751b2-7d8e-4caa-ad61-d8544a501c6b" containerName="console" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.751851 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.768398 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fkb9d"] Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.903483 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-catalog-content\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.903611 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-utilities\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:39 crc kubenswrapper[4710]: I1002 13:01:39.903634 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mnkg\" (UniqueName: \"kubernetes.io/projected/7252945d-16ec-442f-8d1f-a2de5c90bd99-kube-api-access-7mnkg\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.004518 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-catalog-content\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.004822 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-utilities\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.004860 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mnkg\" (UniqueName: \"kubernetes.io/projected/7252945d-16ec-442f-8d1f-a2de5c90bd99-kube-api-access-7mnkg\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.005283 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-utilities\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.005867 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-catalog-content\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.027411 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mnkg\" (UniqueName: \"kubernetes.io/projected/7252945d-16ec-442f-8d1f-a2de5c90bd99-kube-api-access-7mnkg\") pod \"redhat-operators-fkb9d\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.091961 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.428692 4710 generic.go:334] "Generic (PLEG): container finished" podID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerID="31c08a60b02502eb09f9d5870137b8dc13b44ae2328d470b47d54514c0e15bb1" exitCode=0 Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.428782 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" event={"ID":"1675361e-20c2-46ab-8d3a-0b28c519fc3d","Type":"ContainerDied","Data":"31c08a60b02502eb09f9d5870137b8dc13b44ae2328d470b47d54514c0e15bb1"} Oct 02 13:01:40 crc kubenswrapper[4710]: I1002 13:01:40.511677 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fkb9d"] Oct 02 13:01:40 crc kubenswrapper[4710]: W1002 13:01:40.522493 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7252945d_16ec_442f_8d1f_a2de5c90bd99.slice/crio-6a18ae695dc5c867f7a527270dacb0d5b0cead2e884722bc5374367a560c4a3e WatchSource:0}: Error finding container 6a18ae695dc5c867f7a527270dacb0d5b0cead2e884722bc5374367a560c4a3e: Status 404 returned error can't find the container with id 6a18ae695dc5c867f7a527270dacb0d5b0cead2e884722bc5374367a560c4a3e Oct 02 13:01:41 crc kubenswrapper[4710]: I1002 13:01:41.438451 4710 generic.go:334] "Generic (PLEG): container finished" podID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerID="a9f3d5a087b0c4e57d3ba6ddd7b017afa1add2569e589515b404db1486484767" exitCode=0 Oct 02 13:01:41 crc kubenswrapper[4710]: I1002 13:01:41.438572 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" event={"ID":"1675361e-20c2-46ab-8d3a-0b28c519fc3d","Type":"ContainerDied","Data":"a9f3d5a087b0c4e57d3ba6ddd7b017afa1add2569e589515b404db1486484767"} Oct 02 13:01:41 crc kubenswrapper[4710]: I1002 13:01:41.440576 4710 generic.go:334] "Generic (PLEG): container finished" podID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerID="bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9" exitCode=0 Oct 02 13:01:41 crc kubenswrapper[4710]: I1002 13:01:41.440643 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkb9d" event={"ID":"7252945d-16ec-442f-8d1f-a2de5c90bd99","Type":"ContainerDied","Data":"bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9"} Oct 02 13:01:41 crc kubenswrapper[4710]: I1002 13:01:41.440702 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkb9d" event={"ID":"7252945d-16ec-442f-8d1f-a2de5c90bd99","Type":"ContainerStarted","Data":"6a18ae695dc5c867f7a527270dacb0d5b0cead2e884722bc5374367a560c4a3e"} Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.742948 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.754530 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-bundle\") pod \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.754624 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scm6p\" (UniqueName: \"kubernetes.io/projected/1675361e-20c2-46ab-8d3a-0b28c519fc3d-kube-api-access-scm6p\") pod \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.754789 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-util\") pod \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\" (UID: \"1675361e-20c2-46ab-8d3a-0b28c519fc3d\") " Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.760563 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-bundle" (OuterVolumeSpecName: "bundle") pod "1675361e-20c2-46ab-8d3a-0b28c519fc3d" (UID: "1675361e-20c2-46ab-8d3a-0b28c519fc3d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.768971 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1675361e-20c2-46ab-8d3a-0b28c519fc3d-kube-api-access-scm6p" (OuterVolumeSpecName: "kube-api-access-scm6p") pod "1675361e-20c2-46ab-8d3a-0b28c519fc3d" (UID: "1675361e-20c2-46ab-8d3a-0b28c519fc3d"). InnerVolumeSpecName "kube-api-access-scm6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.773621 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-util" (OuterVolumeSpecName: "util") pod "1675361e-20c2-46ab-8d3a-0b28c519fc3d" (UID: "1675361e-20c2-46ab-8d3a-0b28c519fc3d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.856220 4710 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-util\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.856563 4710 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1675361e-20c2-46ab-8d3a-0b28c519fc3d-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:42 crc kubenswrapper[4710]: I1002 13:01:42.856577 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scm6p\" (UniqueName: \"kubernetes.io/projected/1675361e-20c2-46ab-8d3a-0b28c519fc3d-kube-api-access-scm6p\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:43 crc kubenswrapper[4710]: I1002 13:01:43.454832 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" event={"ID":"1675361e-20c2-46ab-8d3a-0b28c519fc3d","Type":"ContainerDied","Data":"cfa4281b82d48105cdef651ab273240f3ee57d28d9a86bac4bcd00f9776fa345"} Oct 02 13:01:43 crc kubenswrapper[4710]: I1002 13:01:43.454876 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfa4281b82d48105cdef651ab273240f3ee57d28d9a86bac4bcd00f9776fa345" Oct 02 13:01:43 crc kubenswrapper[4710]: I1002 13:01:43.454951 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8" Oct 02 13:01:43 crc kubenswrapper[4710]: I1002 13:01:43.456855 4710 generic.go:334] "Generic (PLEG): container finished" podID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerID="e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a" exitCode=0 Oct 02 13:01:43 crc kubenswrapper[4710]: I1002 13:01:43.456920 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkb9d" event={"ID":"7252945d-16ec-442f-8d1f-a2de5c90bd99","Type":"ContainerDied","Data":"e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a"} Oct 02 13:01:45 crc kubenswrapper[4710]: I1002 13:01:45.476082 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkb9d" event={"ID":"7252945d-16ec-442f-8d1f-a2de5c90bd99","Type":"ContainerStarted","Data":"dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d"} Oct 02 13:01:45 crc kubenswrapper[4710]: I1002 13:01:45.505627 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fkb9d" podStartSLOduration=3.191617523 podStartE2EDuration="6.505606211s" podCreationTimestamp="2025-10-02 13:01:39 +0000 UTC" firstStartedPulling="2025-10-02 13:01:41.442407494 +0000 UTC m=+745.548818387" lastFinishedPulling="2025-10-02 13:01:44.756396152 +0000 UTC m=+748.862807075" observedRunningTime="2025-10-02 13:01:45.500111721 +0000 UTC m=+749.606522654" watchObservedRunningTime="2025-10-02 13:01:45.505606211 +0000 UTC m=+749.612017094" Oct 02 13:01:50 crc kubenswrapper[4710]: I1002 13:01:50.092379 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:50 crc kubenswrapper[4710]: I1002 13:01:50.092922 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:50 crc kubenswrapper[4710]: I1002 13:01:50.135549 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:50 crc kubenswrapper[4710]: I1002 13:01:50.548699 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:51 crc kubenswrapper[4710]: I1002 13:01:51.932729 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fkb9d"] Oct 02 13:01:52 crc kubenswrapper[4710]: I1002 13:01:52.518968 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fkb9d" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="registry-server" containerID="cri-o://dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d" gracePeriod=2 Oct 02 13:01:52 crc kubenswrapper[4710]: I1002 13:01:52.530868 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:01:52 crc kubenswrapper[4710]: I1002 13:01:52.530937 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:01:52 crc kubenswrapper[4710]: I1002 13:01:52.989931 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.184097 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-utilities\") pod \"7252945d-16ec-442f-8d1f-a2de5c90bd99\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.184355 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mnkg\" (UniqueName: \"kubernetes.io/projected/7252945d-16ec-442f-8d1f-a2de5c90bd99-kube-api-access-7mnkg\") pod \"7252945d-16ec-442f-8d1f-a2de5c90bd99\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.184399 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-catalog-content\") pod \"7252945d-16ec-442f-8d1f-a2de5c90bd99\" (UID: \"7252945d-16ec-442f-8d1f-a2de5c90bd99\") " Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.185140 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-utilities" (OuterVolumeSpecName: "utilities") pod "7252945d-16ec-442f-8d1f-a2de5c90bd99" (UID: "7252945d-16ec-442f-8d1f-a2de5c90bd99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.205103 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7252945d-16ec-442f-8d1f-a2de5c90bd99-kube-api-access-7mnkg" (OuterVolumeSpecName: "kube-api-access-7mnkg") pod "7252945d-16ec-442f-8d1f-a2de5c90bd99" (UID: "7252945d-16ec-442f-8d1f-a2de5c90bd99"). InnerVolumeSpecName "kube-api-access-7mnkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281359 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7667858fb4-55crv"] Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.281653 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="extract-content" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281674 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="extract-content" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.281686 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="registry-server" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281694 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="registry-server" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.281711 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="pull" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281719 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="pull" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.281733 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="util" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281740 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="util" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.281752 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="extract" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281784 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="extract" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.281800 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="extract-utilities" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281808 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="extract-utilities" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281937 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="1675361e-20c2-46ab-8d3a-0b28c519fc3d" containerName="extract" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.281951 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerName="registry-server" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.282458 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.285892 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.285937 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mnkg\" (UniqueName: \"kubernetes.io/projected/7252945d-16ec-442f-8d1f-a2de5c90bd99-kube-api-access-7mnkg\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.288134 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-kkklr" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.288314 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.288374 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.288468 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.288150 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.299630 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7667858fb4-55crv"] Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.387077 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl4qj\" (UniqueName: \"kubernetes.io/projected/eaac425c-61dd-47c7-a02a-56deb08ca106-kube-api-access-dl4qj\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.387140 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eaac425c-61dd-47c7-a02a-56deb08ca106-apiservice-cert\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.387221 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eaac425c-61dd-47c7-a02a-56deb08ca106-webhook-cert\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.458328 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw"] Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.459077 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.460471 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-srh59" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.460699 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.460961 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.483614 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw"] Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.488619 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eaac425c-61dd-47c7-a02a-56deb08ca106-webhook-cert\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.488695 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl4qj\" (UniqueName: \"kubernetes.io/projected/eaac425c-61dd-47c7-a02a-56deb08ca106-kube-api-access-dl4qj\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.488726 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eaac425c-61dd-47c7-a02a-56deb08ca106-apiservice-cert\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.493080 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eaac425c-61dd-47c7-a02a-56deb08ca106-apiservice-cert\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.503656 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eaac425c-61dd-47c7-a02a-56deb08ca106-webhook-cert\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.515652 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl4qj\" (UniqueName: \"kubernetes.io/projected/eaac425c-61dd-47c7-a02a-56deb08ca106-kube-api-access-dl4qj\") pod \"metallb-operator-controller-manager-7667858fb4-55crv\" (UID: \"eaac425c-61dd-47c7-a02a-56deb08ca106\") " pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.527098 4710 generic.go:334] "Generic (PLEG): container finished" podID="7252945d-16ec-442f-8d1f-a2de5c90bd99" containerID="dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d" exitCode=0 Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.527147 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkb9d" event={"ID":"7252945d-16ec-442f-8d1f-a2de5c90bd99","Type":"ContainerDied","Data":"dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d"} Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.527425 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkb9d" event={"ID":"7252945d-16ec-442f-8d1f-a2de5c90bd99","Type":"ContainerDied","Data":"6a18ae695dc5c867f7a527270dacb0d5b0cead2e884722bc5374367a560c4a3e"} Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.527216 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkb9d" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.527491 4710 scope.go:117] "RemoveContainer" containerID="dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.550141 4710 scope.go:117] "RemoveContainer" containerID="e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.564605 4710 scope.go:117] "RemoveContainer" containerID="bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.580672 4710 scope.go:117] "RemoveContainer" containerID="dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.581077 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d\": container with ID starting with dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d not found: ID does not exist" containerID="dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.581103 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d"} err="failed to get container status \"dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d\": rpc error: code = NotFound desc = could not find container \"dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d\": container with ID starting with dcc8f6fe7911f664758e7793c09a8c09cc55d5683d8bdf9a6dfc5475b8bbd94d not found: ID does not exist" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.581124 4710 scope.go:117] "RemoveContainer" containerID="e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.581285 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a\": container with ID starting with e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a not found: ID does not exist" containerID="e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.581302 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a"} err="failed to get container status \"e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a\": rpc error: code = NotFound desc = could not find container \"e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a\": container with ID starting with e45a13d4696fe151beee389e3a46bc70cc556649d67c0977178dc924b74dbe4a not found: ID does not exist" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.581314 4710 scope.go:117] "RemoveContainer" containerID="bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9" Oct 02 13:01:53 crc kubenswrapper[4710]: E1002 13:01:53.581471 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9\": container with ID starting with bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9 not found: ID does not exist" containerID="bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.581487 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9"} err="failed to get container status \"bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9\": rpc error: code = NotFound desc = could not find container \"bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9\": container with ID starting with bd07389a491d1f8f59c285b15fea289555bf8ba678a0f0cad54ee37993e229a9 not found: ID does not exist" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.590357 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-webhook-cert\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.590430 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-apiservice-cert\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.590461 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sg5d\" (UniqueName: \"kubernetes.io/projected/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-kube-api-access-2sg5d\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.603255 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.691534 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-webhook-cert\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.691892 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-apiservice-cert\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.691917 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sg5d\" (UniqueName: \"kubernetes.io/projected/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-kube-api-access-2sg5d\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.700194 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-apiservice-cert\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.707089 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-webhook-cert\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.711201 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7252945d-16ec-442f-8d1f-a2de5c90bd99" (UID: "7252945d-16ec-442f-8d1f-a2de5c90bd99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.731274 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sg5d\" (UniqueName: \"kubernetes.io/projected/f0d936b3-7f41-4d19-8af8-18b4aa2f8567-kube-api-access-2sg5d\") pod \"metallb-operator-webhook-server-68d8677cb8-7smlw\" (UID: \"f0d936b3-7f41-4d19-8af8-18b4aa2f8567\") " pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.772904 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.793071 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7252945d-16ec-442f-8d1f-a2de5c90bd99-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.870978 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fkb9d"] Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.879123 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fkb9d"] Oct 02 13:01:53 crc kubenswrapper[4710]: I1002 13:01:53.893492 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7667858fb4-55crv"] Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.136567 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-45p6g"] Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.138206 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.156032 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45p6g"] Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.246113 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw"] Oct 02 13:01:54 crc kubenswrapper[4710]: W1002 13:01:54.246688 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0d936b3_7f41_4d19_8af8_18b4aa2f8567.slice/crio-39dc3ad3dae1df7d3db63e7a33e8d1da946d9e7a070e7a66ffe1e030dd572454 WatchSource:0}: Error finding container 39dc3ad3dae1df7d3db63e7a33e8d1da946d9e7a070e7a66ffe1e030dd572454: Status 404 returned error can't find the container with id 39dc3ad3dae1df7d3db63e7a33e8d1da946d9e7a070e7a66ffe1e030dd572454 Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.298508 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrmmf\" (UniqueName: \"kubernetes.io/projected/027171ea-f068-49bb-9662-12e47a7b776e-kube-api-access-lrmmf\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.298850 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-catalog-content\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.298876 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-utilities\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.400268 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrmmf\" (UniqueName: \"kubernetes.io/projected/027171ea-f068-49bb-9662-12e47a7b776e-kube-api-access-lrmmf\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.400366 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-utilities\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.400385 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-catalog-content\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.400861 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-catalog-content\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.400953 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-utilities\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.427530 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrmmf\" (UniqueName: \"kubernetes.io/projected/027171ea-f068-49bb-9662-12e47a7b776e-kube-api-access-lrmmf\") pod \"redhat-marketplace-45p6g\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.452275 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.544267 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" event={"ID":"eaac425c-61dd-47c7-a02a-56deb08ca106","Type":"ContainerStarted","Data":"3b2aa34ffb8b1586457fda5f65c2be6e54f017defc8efa55d1ee5305e98e135c"} Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.546196 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" event={"ID":"f0d936b3-7f41-4d19-8af8-18b4aa2f8567","Type":"ContainerStarted","Data":"39dc3ad3dae1df7d3db63e7a33e8d1da946d9e7a070e7a66ffe1e030dd572454"} Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.870431 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45p6g"] Oct 02 13:01:54 crc kubenswrapper[4710]: I1002 13:01:54.918474 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7252945d-16ec-442f-8d1f-a2de5c90bd99" path="/var/lib/kubelet/pods/7252945d-16ec-442f-8d1f-a2de5c90bd99/volumes" Oct 02 13:01:55 crc kubenswrapper[4710]: I1002 13:01:55.559081 4710 generic.go:334] "Generic (PLEG): container finished" podID="027171ea-f068-49bb-9662-12e47a7b776e" containerID="8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182" exitCode=0 Oct 02 13:01:55 crc kubenswrapper[4710]: I1002 13:01:55.559125 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45p6g" event={"ID":"027171ea-f068-49bb-9662-12e47a7b776e","Type":"ContainerDied","Data":"8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182"} Oct 02 13:01:55 crc kubenswrapper[4710]: I1002 13:01:55.559153 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45p6g" event={"ID":"027171ea-f068-49bb-9662-12e47a7b776e","Type":"ContainerStarted","Data":"e5ec496e9918bb4ce8eaa65209e920e9556381b83a6dd8e5309cc2d04d39c06e"} Oct 02 13:01:56 crc kubenswrapper[4710]: I1002 13:01:56.620778 4710 generic.go:334] "Generic (PLEG): container finished" podID="027171ea-f068-49bb-9662-12e47a7b776e" containerID="a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79" exitCode=0 Oct 02 13:01:56 crc kubenswrapper[4710]: I1002 13:01:56.621333 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45p6g" event={"ID":"027171ea-f068-49bb-9662-12e47a7b776e","Type":"ContainerDied","Data":"a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79"} Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.659534 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" event={"ID":"eaac425c-61dd-47c7-a02a-56deb08ca106","Type":"ContainerStarted","Data":"2e496f71b3aa9777975a57edbf59ededefd4795c6f9d28853348152ed5b98354"} Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.660167 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.663626 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45p6g" event={"ID":"027171ea-f068-49bb-9662-12e47a7b776e","Type":"ContainerStarted","Data":"443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19"} Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.668449 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" event={"ID":"f0d936b3-7f41-4d19-8af8-18b4aa2f8567","Type":"ContainerStarted","Data":"e88c76dfe3e77e3dfb9f75fd40fae3b3b3ca57b0b67c0a5547bc697b061d3483"} Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.668947 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.684913 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" podStartSLOduration=1.667276803 podStartE2EDuration="6.684890041s" podCreationTimestamp="2025-10-02 13:01:53 +0000 UTC" firstStartedPulling="2025-10-02 13:01:53.914912439 +0000 UTC m=+758.021323322" lastFinishedPulling="2025-10-02 13:01:58.932525667 +0000 UTC m=+763.038936560" observedRunningTime="2025-10-02 13:01:59.676897942 +0000 UTC m=+763.783308865" watchObservedRunningTime="2025-10-02 13:01:59.684890041 +0000 UTC m=+763.791300944" Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.700797 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-45p6g" podStartSLOduration=2.329270191 podStartE2EDuration="5.700780686s" podCreationTimestamp="2025-10-02 13:01:54 +0000 UTC" firstStartedPulling="2025-10-02 13:01:55.561101114 +0000 UTC m=+759.667511997" lastFinishedPulling="2025-10-02 13:01:58.932611619 +0000 UTC m=+763.039022492" observedRunningTime="2025-10-02 13:01:59.700359106 +0000 UTC m=+763.806770009" watchObservedRunningTime="2025-10-02 13:01:59.700780686 +0000 UTC m=+763.807191579" Oct 02 13:01:59 crc kubenswrapper[4710]: I1002 13:01:59.730188 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" podStartSLOduration=2.034488138 podStartE2EDuration="6.730167729s" podCreationTimestamp="2025-10-02 13:01:53 +0000 UTC" firstStartedPulling="2025-10-02 13:01:54.250008927 +0000 UTC m=+758.356419810" lastFinishedPulling="2025-10-02 13:01:58.945688518 +0000 UTC m=+763.052099401" observedRunningTime="2025-10-02 13:01:59.72513939 +0000 UTC m=+763.831550283" watchObservedRunningTime="2025-10-02 13:01:59.730167729 +0000 UTC m=+763.836578612" Oct 02 13:02:04 crc kubenswrapper[4710]: I1002 13:02:04.452960 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:02:04 crc kubenswrapper[4710]: I1002 13:02:04.453264 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:02:04 crc kubenswrapper[4710]: I1002 13:02:04.525789 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:02:04 crc kubenswrapper[4710]: I1002 13:02:04.744937 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:02:06 crc kubenswrapper[4710]: I1002 13:02:06.930989 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45p6g"] Oct 02 13:02:06 crc kubenswrapper[4710]: I1002 13:02:06.933251 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-45p6g" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="registry-server" containerID="cri-o://443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19" gracePeriod=2 Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.402995 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.501401 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-catalog-content\") pod \"027171ea-f068-49bb-9662-12e47a7b776e\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.501530 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-utilities\") pod \"027171ea-f068-49bb-9662-12e47a7b776e\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.501601 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrmmf\" (UniqueName: \"kubernetes.io/projected/027171ea-f068-49bb-9662-12e47a7b776e-kube-api-access-lrmmf\") pod \"027171ea-f068-49bb-9662-12e47a7b776e\" (UID: \"027171ea-f068-49bb-9662-12e47a7b776e\") " Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.503665 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-utilities" (OuterVolumeSpecName: "utilities") pod "027171ea-f068-49bb-9662-12e47a7b776e" (UID: "027171ea-f068-49bb-9662-12e47a7b776e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.510468 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/027171ea-f068-49bb-9662-12e47a7b776e-kube-api-access-lrmmf" (OuterVolumeSpecName: "kube-api-access-lrmmf") pod "027171ea-f068-49bb-9662-12e47a7b776e" (UID: "027171ea-f068-49bb-9662-12e47a7b776e"). InnerVolumeSpecName "kube-api-access-lrmmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.516619 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "027171ea-f068-49bb-9662-12e47a7b776e" (UID: "027171ea-f068-49bb-9662-12e47a7b776e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.603503 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.603541 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027171ea-f068-49bb-9662-12e47a7b776e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.603553 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrmmf\" (UniqueName: \"kubernetes.io/projected/027171ea-f068-49bb-9662-12e47a7b776e-kube-api-access-lrmmf\") on node \"crc\" DevicePath \"\"" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.718264 4710 generic.go:334] "Generic (PLEG): container finished" podID="027171ea-f068-49bb-9662-12e47a7b776e" containerID="443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19" exitCode=0 Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.718345 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45p6g" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.718344 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45p6g" event={"ID":"027171ea-f068-49bb-9662-12e47a7b776e","Type":"ContainerDied","Data":"443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19"} Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.718823 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45p6g" event={"ID":"027171ea-f068-49bb-9662-12e47a7b776e","Type":"ContainerDied","Data":"e5ec496e9918bb4ce8eaa65209e920e9556381b83a6dd8e5309cc2d04d39c06e"} Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.718855 4710 scope.go:117] "RemoveContainer" containerID="443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.737270 4710 scope.go:117] "RemoveContainer" containerID="a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.747341 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45p6g"] Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.755896 4710 scope.go:117] "RemoveContainer" containerID="8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.766558 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-45p6g"] Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.780897 4710 scope.go:117] "RemoveContainer" containerID="443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19" Oct 02 13:02:07 crc kubenswrapper[4710]: E1002 13:02:07.781363 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19\": container with ID starting with 443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19 not found: ID does not exist" containerID="443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.781394 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19"} err="failed to get container status \"443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19\": rpc error: code = NotFound desc = could not find container \"443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19\": container with ID starting with 443017e7531bf996e1a1742b624da2e2557f2f063a9877416324ea8b86ccfd19 not found: ID does not exist" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.781414 4710 scope.go:117] "RemoveContainer" containerID="a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79" Oct 02 13:02:07 crc kubenswrapper[4710]: E1002 13:02:07.781728 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79\": container with ID starting with a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79 not found: ID does not exist" containerID="a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.781766 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79"} err="failed to get container status \"a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79\": rpc error: code = NotFound desc = could not find container \"a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79\": container with ID starting with a66a94942ddbf566c9709d42367f44df6a7fc11fb4be29d2f0947dab549cdf79 not found: ID does not exist" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.781779 4710 scope.go:117] "RemoveContainer" containerID="8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182" Oct 02 13:02:07 crc kubenswrapper[4710]: E1002 13:02:07.782018 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182\": container with ID starting with 8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182 not found: ID does not exist" containerID="8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182" Oct 02 13:02:07 crc kubenswrapper[4710]: I1002 13:02:07.782063 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182"} err="failed to get container status \"8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182\": rpc error: code = NotFound desc = could not find container \"8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182\": container with ID starting with 8c742e42ebde4d43684b1a45629a82010d456fba10c9603e6e42137696752182 not found: ID does not exist" Oct 02 13:02:08 crc kubenswrapper[4710]: I1002 13:02:08.911400 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="027171ea-f068-49bb-9662-12e47a7b776e" path="/var/lib/kubelet/pods/027171ea-f068-49bb-9662-12e47a7b776e/volumes" Oct 02 13:02:13 crc kubenswrapper[4710]: I1002 13:02:13.777605 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68d8677cb8-7smlw" Oct 02 13:02:22 crc kubenswrapper[4710]: I1002 13:02:22.531043 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:02:22 crc kubenswrapper[4710]: I1002 13:02:22.531917 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:02:33 crc kubenswrapper[4710]: I1002 13:02:33.607087 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7667858fb4-55crv" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.311104 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-swzvg"] Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.311319 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="extract-utilities" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.311331 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="extract-utilities" Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.311350 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="extract-content" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.311358 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="extract-content" Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.311369 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="registry-server" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.311377 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="registry-server" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.311515 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="027171ea-f068-49bb-9662-12e47a7b776e" containerName="registry-server" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.313524 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.315384 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.315524 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-sjp4b" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.318375 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.321019 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2"] Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.321995 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.323540 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.332313 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2"] Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.379895 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-reloader\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.379955 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x92wl\" (UniqueName: \"kubernetes.io/projected/bfc4a046-7c49-44e9-8324-3d5b54c57442-kube-api-access-x92wl\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380012 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-startup\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380093 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-metrics\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380151 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ac0bab7-e1c0-4411-8a35-18fa210385a7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380227 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-sockets\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380344 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-conf\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380396 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhhcr\" (UniqueName: \"kubernetes.io/projected/6ac0bab7-e1c0-4411-8a35-18fa210385a7-kube-api-access-fhhcr\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.380491 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfc4a046-7c49-44e9-8324-3d5b54c57442-metrics-certs\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.386787 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-wh9nb"] Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.387683 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.389813 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bktwk" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.390045 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.390097 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.390226 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.410413 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-4x985"] Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.411261 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: W1002 13:02:34.424082 4710 reflector.go:561] object-"metallb-system"/"controller-certs-secret": failed to list *v1.Secret: secrets "controller-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.424134 4710 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"controller-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"controller-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.446248 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-4x985"] Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.481764 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjwjw\" (UniqueName: \"kubernetes.io/projected/73bea163-fba2-4a76-8e88-6a31ed86f8f3-kube-api-access-fjwjw\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.481833 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-metrics\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.481867 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ac0bab7-e1c0-4411-8a35-18fa210385a7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.481908 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-sockets\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.481962 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-conf\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.482012 4710 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.482085 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ac0bab7-e1c0-4411-8a35-18fa210385a7-cert podName:6ac0bab7-e1c0-4411-8a35-18fa210385a7 nodeName:}" failed. No retries permitted until 2025-10-02 13:02:34.982067485 +0000 UTC m=+799.088478358 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6ac0bab7-e1c0-4411-8a35-18fa210385a7-cert") pod "frr-k8s-webhook-server-64bf5d555-wv5f2" (UID: "6ac0bab7-e1c0-4411-8a35-18fa210385a7") : secret "frr-k8s-webhook-server-cert" not found Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.482428 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-metrics\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.482472 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-sockets\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.482546 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhhcr\" (UniqueName: \"kubernetes.io/projected/6ac0bab7-e1c0-4411-8a35-18fa210385a7-kube-api-access-fhhcr\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.482583 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-conf\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.482904 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metrics-certs\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.482958 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metallb-excludel2\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.483033 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfc4a046-7c49-44e9-8324-3d5b54c57442-metrics-certs\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.483060 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-metrics-certs\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.483980 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-reloader\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.484026 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x92wl\" (UniqueName: \"kubernetes.io/projected/bfc4a046-7c49-44e9-8324-3d5b54c57442-kube-api-access-x92wl\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.484053 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6wbb\" (UniqueName: \"kubernetes.io/projected/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-kube-api-access-p6wbb\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.484074 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.484100 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-startup\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.484137 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-cert\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.484405 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bfc4a046-7c49-44e9-8324-3d5b54c57442-reloader\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.485456 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bfc4a046-7c49-44e9-8324-3d5b54c57442-frr-startup\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.499831 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfc4a046-7c49-44e9-8324-3d5b54c57442-metrics-certs\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.504977 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhhcr\" (UniqueName: \"kubernetes.io/projected/6ac0bab7-e1c0-4411-8a35-18fa210385a7-kube-api-access-fhhcr\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.505351 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x92wl\" (UniqueName: \"kubernetes.io/projected/bfc4a046-7c49-44e9-8324-3d5b54c57442-kube-api-access-x92wl\") pod \"frr-k8s-swzvg\" (UID: \"bfc4a046-7c49-44e9-8324-3d5b54c57442\") " pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585624 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metrics-certs\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585671 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metallb-excludel2\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585702 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-metrics-certs\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585760 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6wbb\" (UniqueName: \"kubernetes.io/projected/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-kube-api-access-p6wbb\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585782 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585805 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-cert\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.585829 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjwjw\" (UniqueName: \"kubernetes.io/projected/73bea163-fba2-4a76-8e88-6a31ed86f8f3-kube-api-access-fjwjw\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.585904 4710 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.585997 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metrics-certs podName:73bea163-fba2-4a76-8e88-6a31ed86f8f3 nodeName:}" failed. No retries permitted until 2025-10-02 13:02:35.085974332 +0000 UTC m=+799.192385285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metrics-certs") pod "speaker-wh9nb" (UID: "73bea163-fba2-4a76-8e88-6a31ed86f8f3") : secret "speaker-certs-secret" not found Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.585995 4710 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 13:02:34 crc kubenswrapper[4710]: E1002 13:02:34.586115 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist podName:73bea163-fba2-4a76-8e88-6a31ed86f8f3 nodeName:}" failed. No retries permitted until 2025-10-02 13:02:35.086090185 +0000 UTC m=+799.192501168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist") pod "speaker-wh9nb" (UID: "73bea163-fba2-4a76-8e88-6a31ed86f8f3") : secret "metallb-memberlist" not found Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.586823 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metallb-excludel2\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.591506 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.599922 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-cert\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.608059 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjwjw\" (UniqueName: \"kubernetes.io/projected/73bea163-fba2-4a76-8e88-6a31ed86f8f3-kube-api-access-fjwjw\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.617036 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6wbb\" (UniqueName: \"kubernetes.io/projected/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-kube-api-access-p6wbb\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.631406 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.883947 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"3c2f15afb2e49a3b5b79f7c4c05b385e43c1664ce0f1e896c6375b9fd8dba371"} Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.991138 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ac0bab7-e1c0-4411-8a35-18fa210385a7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:34 crc kubenswrapper[4710]: I1002 13:02:34.996832 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ac0bab7-e1c0-4411-8a35-18fa210385a7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wv5f2\" (UID: \"6ac0bab7-e1c0-4411-8a35-18fa210385a7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.092057 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metrics-certs\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.092167 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:35 crc kubenswrapper[4710]: E1002 13:02:35.092344 4710 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 13:02:35 crc kubenswrapper[4710]: E1002 13:02:35.092408 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist podName:73bea163-fba2-4a76-8e88-6a31ed86f8f3 nodeName:}" failed. No retries permitted until 2025-10-02 13:02:36.092387046 +0000 UTC m=+800.198797939 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist") pod "speaker-wh9nb" (UID: "73bea163-fba2-4a76-8e88-6a31ed86f8f3") : secret "metallb-memberlist" not found Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.097829 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-metrics-certs\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.240986 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.378100 4710 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.389869 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b1c2aba-9eb9-4851-9bbe-75c4ee93f325-metrics-certs\") pod \"controller-68d546b9d8-4x985\" (UID: \"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325\") " pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.624334 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.642669 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2"] Oct 02 13:02:35 crc kubenswrapper[4710]: W1002 13:02:35.661989 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ac0bab7_e1c0_4411_8a35_18fa210385a7.slice/crio-c9c8e99ac49249b69eb9a1db572efa7fd495138ef182a52859be48defc5014ed WatchSource:0}: Error finding container c9c8e99ac49249b69eb9a1db572efa7fd495138ef182a52859be48defc5014ed: Status 404 returned error can't find the container with id c9c8e99ac49249b69eb9a1db572efa7fd495138ef182a52859be48defc5014ed Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.855238 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-4x985"] Oct 02 13:02:35 crc kubenswrapper[4710]: W1002 13:02:35.865051 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1c2aba_9eb9_4851_9bbe_75c4ee93f325.slice/crio-8ad8f520f50394317646b88aa3ad30d00712718252fb800246e192be13a40fbf WatchSource:0}: Error finding container 8ad8f520f50394317646b88aa3ad30d00712718252fb800246e192be13a40fbf: Status 404 returned error can't find the container with id 8ad8f520f50394317646b88aa3ad30d00712718252fb800246e192be13a40fbf Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.890302 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" event={"ID":"6ac0bab7-e1c0-4411-8a35-18fa210385a7","Type":"ContainerStarted","Data":"c9c8e99ac49249b69eb9a1db572efa7fd495138ef182a52859be48defc5014ed"} Oct 02 13:02:35 crc kubenswrapper[4710]: I1002 13:02:35.891282 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4x985" event={"ID":"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325","Type":"ContainerStarted","Data":"8ad8f520f50394317646b88aa3ad30d00712718252fb800246e192be13a40fbf"} Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.110238 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.114993 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/73bea163-fba2-4a76-8e88-6a31ed86f8f3-memberlist\") pod \"speaker-wh9nb\" (UID: \"73bea163-fba2-4a76-8e88-6a31ed86f8f3\") " pod="metallb-system/speaker-wh9nb" Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.201262 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wh9nb" Oct 02 13:02:36 crc kubenswrapper[4710]: W1002 13:02:36.217277 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73bea163_fba2_4a76_8e88_6a31ed86f8f3.slice/crio-92470d51a26736622e59c2560318e2a07834f55a9fefcd253a09b8e4ffa9cd7e WatchSource:0}: Error finding container 92470d51a26736622e59c2560318e2a07834f55a9fefcd253a09b8e4ffa9cd7e: Status 404 returned error can't find the container with id 92470d51a26736622e59c2560318e2a07834f55a9fefcd253a09b8e4ffa9cd7e Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.899569 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4x985" event={"ID":"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325","Type":"ContainerStarted","Data":"ab95d9ed4714d5e882331dcf54a4e74751f83fd7e467646ae836be20e4cf02ed"} Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.899998 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4x985" event={"ID":"3b1c2aba-9eb9-4851-9bbe-75c4ee93f325","Type":"ContainerStarted","Data":"37c8c98411b3ab5bac7d35afda316a270737bea266ab64b1492b9b97a85d6ff1"} Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.900033 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.901815 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wh9nb" event={"ID":"73bea163-fba2-4a76-8e88-6a31ed86f8f3","Type":"ContainerStarted","Data":"5da2e08de007f9854c67f4bd04f61298647df32a287cf7a85cb57e78e298ae1b"} Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.901861 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wh9nb" event={"ID":"73bea163-fba2-4a76-8e88-6a31ed86f8f3","Type":"ContainerStarted","Data":"9d6c00373e037afc762748b7fc11bab9101166d599c1e7a0fb6b84e78990b2f3"} Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.901870 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wh9nb" event={"ID":"73bea163-fba2-4a76-8e88-6a31ed86f8f3","Type":"ContainerStarted","Data":"92470d51a26736622e59c2560318e2a07834f55a9fefcd253a09b8e4ffa9cd7e"} Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.902057 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wh9nb" Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.919817 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-4x985" podStartSLOduration=2.919799824 podStartE2EDuration="2.919799824s" podCreationTimestamp="2025-10-02 13:02:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:02:36.916446095 +0000 UTC m=+801.022856998" watchObservedRunningTime="2025-10-02 13:02:36.919799824 +0000 UTC m=+801.026210707" Oct 02 13:02:36 crc kubenswrapper[4710]: I1002 13:02:36.948080 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-wh9nb" podStartSLOduration=2.948056417 podStartE2EDuration="2.948056417s" podCreationTimestamp="2025-10-02 13:02:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:02:36.943429538 +0000 UTC m=+801.049840451" watchObservedRunningTime="2025-10-02 13:02:36.948056417 +0000 UTC m=+801.054467300" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.594663 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7gssz"] Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.597399 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.606924 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gssz"] Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.691024 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rcz2\" (UniqueName: \"kubernetes.io/projected/7e27adf8-1ad4-41f4-a411-5698496f46ea-kube-api-access-5rcz2\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.691113 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-utilities\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.691136 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-catalog-content\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.792142 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rcz2\" (UniqueName: \"kubernetes.io/projected/7e27adf8-1ad4-41f4-a411-5698496f46ea-kube-api-access-5rcz2\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.792246 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-utilities\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.792273 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-catalog-content\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.792820 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-catalog-content\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.792866 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-utilities\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.812877 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rcz2\" (UniqueName: \"kubernetes.io/projected/7e27adf8-1ad4-41f4-a411-5698496f46ea-kube-api-access-5rcz2\") pod \"certified-operators-7gssz\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:41 crc kubenswrapper[4710]: I1002 13:02:41.928337 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:42 crc kubenswrapper[4710]: I1002 13:02:42.944497 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" event={"ID":"6ac0bab7-e1c0-4411-8a35-18fa210385a7","Type":"ContainerStarted","Data":"1ece48ad48de09f1348360f587308be1099a9ea54b69d2e2853d268a9e259d55"} Oct 02 13:02:42 crc kubenswrapper[4710]: I1002 13:02:42.945118 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:42 crc kubenswrapper[4710]: I1002 13:02:42.946079 4710 generic.go:334] "Generic (PLEG): container finished" podID="bfc4a046-7c49-44e9-8324-3d5b54c57442" containerID="9d253a5cb1d5302d194a90077ae625cf5c7180b18222e7276f8fcc40081552f4" exitCode=0 Oct 02 13:02:42 crc kubenswrapper[4710]: I1002 13:02:42.946107 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerDied","Data":"9d253a5cb1d5302d194a90077ae625cf5c7180b18222e7276f8fcc40081552f4"} Oct 02 13:02:42 crc kubenswrapper[4710]: I1002 13:02:42.961256 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" podStartSLOduration=1.968221283 podStartE2EDuration="8.961234882s" podCreationTimestamp="2025-10-02 13:02:34 +0000 UTC" firstStartedPulling="2025-10-02 13:02:35.664486121 +0000 UTC m=+799.770896994" lastFinishedPulling="2025-10-02 13:02:42.65749971 +0000 UTC m=+806.763910593" observedRunningTime="2025-10-02 13:02:42.958665372 +0000 UTC m=+807.065076255" watchObservedRunningTime="2025-10-02 13:02:42.961234882 +0000 UTC m=+807.067645765" Oct 02 13:02:42 crc kubenswrapper[4710]: I1002 13:02:42.980854 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gssz"] Oct 02 13:02:42 crc kubenswrapper[4710]: W1002 13:02:42.989214 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e27adf8_1ad4_41f4_a411_5698496f46ea.slice/crio-f41b6d0cf4cec17c45c271a9e5ee01cd005d24f50f79cf9780414c5c1a6f9323 WatchSource:0}: Error finding container f41b6d0cf4cec17c45c271a9e5ee01cd005d24f50f79cf9780414c5c1a6f9323: Status 404 returned error can't find the container with id f41b6d0cf4cec17c45c271a9e5ee01cd005d24f50f79cf9780414c5c1a6f9323 Oct 02 13:02:43 crc kubenswrapper[4710]: I1002 13:02:43.953294 4710 generic.go:334] "Generic (PLEG): container finished" podID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerID="9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1" exitCode=0 Oct 02 13:02:43 crc kubenswrapper[4710]: I1002 13:02:43.953403 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerDied","Data":"9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1"} Oct 02 13:02:43 crc kubenswrapper[4710]: I1002 13:02:43.953696 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerStarted","Data":"f41b6d0cf4cec17c45c271a9e5ee01cd005d24f50f79cf9780414c5c1a6f9323"} Oct 02 13:02:43 crc kubenswrapper[4710]: I1002 13:02:43.956293 4710 generic.go:334] "Generic (PLEG): container finished" podID="bfc4a046-7c49-44e9-8324-3d5b54c57442" containerID="9edd0d03e2696151e7ec8fde24c82e4a3ce56c68a7d6916bf2ef95949359acb8" exitCode=0 Oct 02 13:02:43 crc kubenswrapper[4710]: I1002 13:02:43.956726 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerDied","Data":"9edd0d03e2696151e7ec8fde24c82e4a3ce56c68a7d6916bf2ef95949359acb8"} Oct 02 13:02:44 crc kubenswrapper[4710]: I1002 13:02:44.965215 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerStarted","Data":"43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330"} Oct 02 13:02:44 crc kubenswrapper[4710]: I1002 13:02:44.969002 4710 generic.go:334] "Generic (PLEG): container finished" podID="bfc4a046-7c49-44e9-8324-3d5b54c57442" containerID="5c93651644053c826df7fd3ce1e69fe123626403c615599f61692fccff72cf5e" exitCode=0 Oct 02 13:02:44 crc kubenswrapper[4710]: I1002 13:02:44.969069 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerDied","Data":"5c93651644053c826df7fd3ce1e69fe123626403c615599f61692fccff72cf5e"} Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.628740 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-4x985" Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.978432 4710 generic.go:334] "Generic (PLEG): container finished" podID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerID="43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330" exitCode=0 Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.978484 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerDied","Data":"43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330"} Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.984535 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"89abbeb5740c49242d59892f3bada841b9beddb3914be4a5aaa8c0f81cb937b1"} Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.984854 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"17b923ea1d257e4db49b24e0d412fbccbe3b682e20375bc305934b8ed051cdb6"} Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.984865 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"e48b4068a012cbf5ddb343cc7ee9d421208e66afa5fada7fd4da113c532cc8f0"} Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.984873 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"3e2bd056dff9cffade5243f03c60286e8edb4719bc2721d8bc1a4b28c63db6f5"} Oct 02 13:02:45 crc kubenswrapper[4710]: I1002 13:02:45.984880 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"68645f47337926045b1a1eb77be7847f2e4c402e6300913d5c10b2864927e1ef"} Oct 02 13:02:46 crc kubenswrapper[4710]: I1002 13:02:46.209889 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wh9nb" Oct 02 13:02:46 crc kubenswrapper[4710]: I1002 13:02:46.996003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerStarted","Data":"5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3"} Oct 02 13:02:47 crc kubenswrapper[4710]: I1002 13:02:47.012814 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-swzvg" event={"ID":"bfc4a046-7c49-44e9-8324-3d5b54c57442","Type":"ContainerStarted","Data":"1c5e5e880f82efa9f749ef283b5b430ef6bddc96191b3294ee9a8e54a3771faa"} Oct 02 13:02:47 crc kubenswrapper[4710]: I1002 13:02:47.013629 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:47 crc kubenswrapper[4710]: I1002 13:02:47.023591 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7gssz" podStartSLOduration=3.490140331 podStartE2EDuration="6.023562904s" podCreationTimestamp="2025-10-02 13:02:41 +0000 UTC" firstStartedPulling="2025-10-02 13:02:43.955462304 +0000 UTC m=+808.061873187" lastFinishedPulling="2025-10-02 13:02:46.488884877 +0000 UTC m=+810.595295760" observedRunningTime="2025-10-02 13:02:47.019285624 +0000 UTC m=+811.125696517" watchObservedRunningTime="2025-10-02 13:02:47.023562904 +0000 UTC m=+811.129973787" Oct 02 13:02:47 crc kubenswrapper[4710]: I1002 13:02:47.051351 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-swzvg" podStartSLOduration=5.126379093 podStartE2EDuration="13.051327905s" podCreationTimestamp="2025-10-02 13:02:34 +0000 UTC" firstStartedPulling="2025-10-02 13:02:34.771603954 +0000 UTC m=+798.878014837" lastFinishedPulling="2025-10-02 13:02:42.696552766 +0000 UTC m=+806.802963649" observedRunningTime="2025-10-02 13:02:47.046859581 +0000 UTC m=+811.153270464" watchObservedRunningTime="2025-10-02 13:02:47.051327905 +0000 UTC m=+811.157738788" Oct 02 13:02:49 crc kubenswrapper[4710]: I1002 13:02:49.633093 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:49 crc kubenswrapper[4710]: I1002 13:02:49.669886 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-swzvg" Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.776489 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-f6g5f"] Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.777362 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.782535 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.782658 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.792907 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-smtrz" Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.815570 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f6g5f"] Oct 02 13:02:50 crc kubenswrapper[4710]: I1002 13:02:50.915531 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnh6d\" (UniqueName: \"kubernetes.io/projected/925e710b-d243-4055-8133-ad8bb3cd81bd-kube-api-access-lnh6d\") pod \"openstack-operator-index-f6g5f\" (UID: \"925e710b-d243-4055-8133-ad8bb3cd81bd\") " pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.017206 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnh6d\" (UniqueName: \"kubernetes.io/projected/925e710b-d243-4055-8133-ad8bb3cd81bd-kube-api-access-lnh6d\") pod \"openstack-operator-index-f6g5f\" (UID: \"925e710b-d243-4055-8133-ad8bb3cd81bd\") " pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.042855 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnh6d\" (UniqueName: \"kubernetes.io/projected/925e710b-d243-4055-8133-ad8bb3cd81bd-kube-api-access-lnh6d\") pod \"openstack-operator-index-f6g5f\" (UID: \"925e710b-d243-4055-8133-ad8bb3cd81bd\") " pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.119730 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.567064 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f6g5f"] Oct 02 13:02:51 crc kubenswrapper[4710]: W1002 13:02:51.574319 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod925e710b_d243_4055_8133_ad8bb3cd81bd.slice/crio-57c0ea91e9108b6119a326c0600b858d8fecc79db38bbbae4cf7fe1f4aec1314 WatchSource:0}: Error finding container 57c0ea91e9108b6119a326c0600b858d8fecc79db38bbbae4cf7fe1f4aec1314: Status 404 returned error can't find the container with id 57c0ea91e9108b6119a326c0600b858d8fecc79db38bbbae4cf7fe1f4aec1314 Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.929414 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.929493 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:51 crc kubenswrapper[4710]: I1002 13:02:51.979727 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.051056 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f6g5f" event={"ID":"925e710b-d243-4055-8133-ad8bb3cd81bd","Type":"ContainerStarted","Data":"57c0ea91e9108b6119a326c0600b858d8fecc79db38bbbae4cf7fe1f4aec1314"} Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.088967 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.530638 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.530707 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.530787 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.531437 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2554f6884f6fea884aeb14ab54b4450d5c8dcfe9721cc87320ad6736f5b0d86e"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:02:52 crc kubenswrapper[4710]: I1002 13:02:52.531534 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://2554f6884f6fea884aeb14ab54b4450d5c8dcfe9721cc87320ad6736f5b0d86e" gracePeriod=600 Oct 02 13:02:53 crc kubenswrapper[4710]: I1002 13:02:53.060648 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="2554f6884f6fea884aeb14ab54b4450d5c8dcfe9721cc87320ad6736f5b0d86e" exitCode=0 Oct 02 13:02:53 crc kubenswrapper[4710]: I1002 13:02:53.060738 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"2554f6884f6fea884aeb14ab54b4450d5c8dcfe9721cc87320ad6736f5b0d86e"} Oct 02 13:02:53 crc kubenswrapper[4710]: I1002 13:02:53.061072 4710 scope.go:117] "RemoveContainer" containerID="579103ebcb45d750aaacc6f76838d855e05a1b531da1b682b820b0d029b5b04d" Oct 02 13:02:54 crc kubenswrapper[4710]: I1002 13:02:54.069135 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"be59f3d65c42b4df4c251f325b3eef5d2ce156af0437d5494544eb88ffa0a1fd"} Oct 02 13:02:54 crc kubenswrapper[4710]: I1002 13:02:54.070422 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f6g5f" event={"ID":"925e710b-d243-4055-8133-ad8bb3cd81bd","Type":"ContainerStarted","Data":"13c177572ea55a713f76507a057f6a838bb1ca0ba22605599456b029220f6dba"} Oct 02 13:02:54 crc kubenswrapper[4710]: I1002 13:02:54.104379 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-f6g5f" podStartSLOduration=1.839593416 podStartE2EDuration="4.104359681s" podCreationTimestamp="2025-10-02 13:02:50 +0000 UTC" firstStartedPulling="2025-10-02 13:02:51.576733943 +0000 UTC m=+815.683144826" lastFinishedPulling="2025-10-02 13:02:53.841500208 +0000 UTC m=+817.947911091" observedRunningTime="2025-10-02 13:02:54.102277242 +0000 UTC m=+818.208688145" watchObservedRunningTime="2025-10-02 13:02:54.104359681 +0000 UTC m=+818.210770564" Oct 02 13:02:55 crc kubenswrapper[4710]: I1002 13:02:55.245848 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wv5f2" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.179398 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xxrq9"] Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.181261 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.190761 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xxrq9"] Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.310340 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-utilities\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.310487 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz5g6\" (UniqueName: \"kubernetes.io/projected/e91312f6-8834-4188-9515-f0767e8e7cd8-kube-api-access-wz5g6\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.310544 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-catalog-content\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.411463 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-catalog-content\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.411544 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-utilities\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.411621 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz5g6\" (UniqueName: \"kubernetes.io/projected/e91312f6-8834-4188-9515-f0767e8e7cd8-kube-api-access-wz5g6\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.412274 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-catalog-content\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.412288 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-utilities\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.430084 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz5g6\" (UniqueName: \"kubernetes.io/projected/e91312f6-8834-4188-9515-f0767e8e7cd8-kube-api-access-wz5g6\") pod \"community-operators-xxrq9\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.511757 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:02:57 crc kubenswrapper[4710]: I1002 13:02:57.936272 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xxrq9"] Oct 02 13:02:57 crc kubenswrapper[4710]: W1002 13:02:57.940864 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode91312f6_8834_4188_9515_f0767e8e7cd8.slice/crio-61b6d949e87941fd2c6997510651c29c9dcfc048e06916b792285f39202b367f WatchSource:0}: Error finding container 61b6d949e87941fd2c6997510651c29c9dcfc048e06916b792285f39202b367f: Status 404 returned error can't find the container with id 61b6d949e87941fd2c6997510651c29c9dcfc048e06916b792285f39202b367f Oct 02 13:02:58 crc kubenswrapper[4710]: I1002 13:02:58.126494 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerStarted","Data":"61b6d949e87941fd2c6997510651c29c9dcfc048e06916b792285f39202b367f"} Oct 02 13:02:59 crc kubenswrapper[4710]: I1002 13:02:59.136813 4710 generic.go:334] "Generic (PLEG): container finished" podID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerID="3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae" exitCode=0 Oct 02 13:02:59 crc kubenswrapper[4710]: I1002 13:02:59.136892 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerDied","Data":"3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae"} Oct 02 13:03:00 crc kubenswrapper[4710]: I1002 13:03:00.147207 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerStarted","Data":"76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f"} Oct 02 13:03:01 crc kubenswrapper[4710]: I1002 13:03:01.120102 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:03:01 crc kubenswrapper[4710]: I1002 13:03:01.120199 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:03:01 crc kubenswrapper[4710]: I1002 13:03:01.161485 4710 generic.go:334] "Generic (PLEG): container finished" podID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerID="76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f" exitCode=0 Oct 02 13:03:01 crc kubenswrapper[4710]: I1002 13:03:01.161575 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerDied","Data":"76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f"} Oct 02 13:03:01 crc kubenswrapper[4710]: I1002 13:03:01.163924 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:03:01 crc kubenswrapper[4710]: I1002 13:03:01.206541 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-f6g5f" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.169858 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerStarted","Data":"7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e"} Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.171310 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gssz"] Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.171488 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7gssz" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="registry-server" containerID="cri-o://5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3" gracePeriod=2 Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.194429 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xxrq9" podStartSLOduration=2.514156687 podStartE2EDuration="5.194412682s" podCreationTimestamp="2025-10-02 13:02:57 +0000 UTC" firstStartedPulling="2025-10-02 13:02:59.138877608 +0000 UTC m=+823.245288491" lastFinishedPulling="2025-10-02 13:03:01.819133593 +0000 UTC m=+825.925544486" observedRunningTime="2025-10-02 13:03:02.189529138 +0000 UTC m=+826.295940041" watchObservedRunningTime="2025-10-02 13:03:02.194412682 +0000 UTC m=+826.300823565" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.562600 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.683342 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rcz2\" (UniqueName: \"kubernetes.io/projected/7e27adf8-1ad4-41f4-a411-5698496f46ea-kube-api-access-5rcz2\") pod \"7e27adf8-1ad4-41f4-a411-5698496f46ea\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.683585 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-catalog-content\") pod \"7e27adf8-1ad4-41f4-a411-5698496f46ea\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.683683 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-utilities\") pod \"7e27adf8-1ad4-41f4-a411-5698496f46ea\" (UID: \"7e27adf8-1ad4-41f4-a411-5698496f46ea\") " Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.684742 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-utilities" (OuterVolumeSpecName: "utilities") pod "7e27adf8-1ad4-41f4-a411-5698496f46ea" (UID: "7e27adf8-1ad4-41f4-a411-5698496f46ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.689192 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e27adf8-1ad4-41f4-a411-5698496f46ea-kube-api-access-5rcz2" (OuterVolumeSpecName: "kube-api-access-5rcz2") pod "7e27adf8-1ad4-41f4-a411-5698496f46ea" (UID: "7e27adf8-1ad4-41f4-a411-5698496f46ea"). InnerVolumeSpecName "kube-api-access-5rcz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.736209 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e27adf8-1ad4-41f4-a411-5698496f46ea" (UID: "7e27adf8-1ad4-41f4-a411-5698496f46ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.785952 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.786000 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e27adf8-1ad4-41f4-a411-5698496f46ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:02 crc kubenswrapper[4710]: I1002 13:03:02.786018 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rcz2\" (UniqueName: \"kubernetes.io/projected/7e27adf8-1ad4-41f4-a411-5698496f46ea-kube-api-access-5rcz2\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.182511 4710 generic.go:334] "Generic (PLEG): container finished" podID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerID="5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3" exitCode=0 Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.182584 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerDied","Data":"5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3"} Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.182897 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gssz" event={"ID":"7e27adf8-1ad4-41f4-a411-5698496f46ea","Type":"ContainerDied","Data":"f41b6d0cf4cec17c45c271a9e5ee01cd005d24f50f79cf9780414c5c1a6f9323"} Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.182935 4710 scope.go:117] "RemoveContainer" containerID="5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.182679 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gssz" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.205823 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gssz"] Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.208164 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7gssz"] Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.213495 4710 scope.go:117] "RemoveContainer" containerID="43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.237894 4710 scope.go:117] "RemoveContainer" containerID="9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.267725 4710 scope.go:117] "RemoveContainer" containerID="5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3" Oct 02 13:03:03 crc kubenswrapper[4710]: E1002 13:03:03.268360 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3\": container with ID starting with 5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3 not found: ID does not exist" containerID="5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.268401 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3"} err="failed to get container status \"5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3\": rpc error: code = NotFound desc = could not find container \"5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3\": container with ID starting with 5abe41f1cd96a319a72e4da370db2448cde84a21de9f123dcd67e081cb80dca3 not found: ID does not exist" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.268427 4710 scope.go:117] "RemoveContainer" containerID="43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330" Oct 02 13:03:03 crc kubenswrapper[4710]: E1002 13:03:03.268666 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330\": container with ID starting with 43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330 not found: ID does not exist" containerID="43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.268686 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330"} err="failed to get container status \"43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330\": rpc error: code = NotFound desc = could not find container \"43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330\": container with ID starting with 43d8b9103f13aa12c499d99d42765c08a38fce13f2e71f5ecc89afc77305b330 not found: ID does not exist" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.268700 4710 scope.go:117] "RemoveContainer" containerID="9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1" Oct 02 13:03:03 crc kubenswrapper[4710]: E1002 13:03:03.269201 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1\": container with ID starting with 9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1 not found: ID does not exist" containerID="9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.269268 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1"} err="failed to get container status \"9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1\": rpc error: code = NotFound desc = could not find container \"9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1\": container with ID starting with 9ab713f1b2462d20d740ee4a3da7e728118b73a56c767aae4945c847858a44e1 not found: ID does not exist" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.410098 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49"] Oct 02 13:03:03 crc kubenswrapper[4710]: E1002 13:03:03.410340 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="extract-utilities" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.410353 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="extract-utilities" Oct 02 13:03:03 crc kubenswrapper[4710]: E1002 13:03:03.410362 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="extract-content" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.410369 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="extract-content" Oct 02 13:03:03 crc kubenswrapper[4710]: E1002 13:03:03.410379 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="registry-server" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.410385 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="registry-server" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.410512 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" containerName="registry-server" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.411326 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.413171 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-c5qzn" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.421406 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49"] Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.494509 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-util\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.494576 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-bundle\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.494613 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d755r\" (UniqueName: \"kubernetes.io/projected/7986fe9f-1265-4339-ac4c-24fba0067f48-kube-api-access-d755r\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.595587 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d755r\" (UniqueName: \"kubernetes.io/projected/7986fe9f-1265-4339-ac4c-24fba0067f48-kube-api-access-d755r\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.595873 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-util\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.596040 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-bundle\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.596404 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-bundle\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.596462 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-util\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.624784 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d755r\" (UniqueName: \"kubernetes.io/projected/7986fe9f-1265-4339-ac4c-24fba0067f48-kube-api-access-d755r\") pod \"9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:03 crc kubenswrapper[4710]: I1002 13:03:03.724802 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:04 crc kubenswrapper[4710]: I1002 13:03:04.185115 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49"] Oct 02 13:03:04 crc kubenswrapper[4710]: W1002 13:03:04.190809 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7986fe9f_1265_4339_ac4c_24fba0067f48.slice/crio-bdf8739845f5a16ab91b064b50cc88d1d526168f7d014038f8d1bf3aa864db18 WatchSource:0}: Error finding container bdf8739845f5a16ab91b064b50cc88d1d526168f7d014038f8d1bf3aa864db18: Status 404 returned error can't find the container with id bdf8739845f5a16ab91b064b50cc88d1d526168f7d014038f8d1bf3aa864db18 Oct 02 13:03:04 crc kubenswrapper[4710]: I1002 13:03:04.635402 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-swzvg" Oct 02 13:03:04 crc kubenswrapper[4710]: I1002 13:03:04.914085 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e27adf8-1ad4-41f4-a411-5698496f46ea" path="/var/lib/kubelet/pods/7e27adf8-1ad4-41f4-a411-5698496f46ea/volumes" Oct 02 13:03:05 crc kubenswrapper[4710]: I1002 13:03:05.204951 4710 generic.go:334] "Generic (PLEG): container finished" podID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerID="7711814bdb6b954c1929bf378a73127cc5a7d3ebc65ff561f7c3a55cfc38ade7" exitCode=0 Oct 02 13:03:05 crc kubenswrapper[4710]: I1002 13:03:05.205028 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" event={"ID":"7986fe9f-1265-4339-ac4c-24fba0067f48","Type":"ContainerDied","Data":"7711814bdb6b954c1929bf378a73127cc5a7d3ebc65ff561f7c3a55cfc38ade7"} Oct 02 13:03:05 crc kubenswrapper[4710]: I1002 13:03:05.205266 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" event={"ID":"7986fe9f-1265-4339-ac4c-24fba0067f48","Type":"ContainerStarted","Data":"bdf8739845f5a16ab91b064b50cc88d1d526168f7d014038f8d1bf3aa864db18"} Oct 02 13:03:07 crc kubenswrapper[4710]: I1002 13:03:07.512849 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:03:07 crc kubenswrapper[4710]: I1002 13:03:07.513160 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:03:07 crc kubenswrapper[4710]: I1002 13:03:07.555758 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:03:08 crc kubenswrapper[4710]: I1002 13:03:08.301708 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:03:09 crc kubenswrapper[4710]: I1002 13:03:09.236432 4710 generic.go:334] "Generic (PLEG): container finished" podID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerID="8d60d7bdcd4d6b4a9e3e572101586f33cb84df92a4f495bdf1dc2e24e0f0dc95" exitCode=0 Oct 02 13:03:09 crc kubenswrapper[4710]: I1002 13:03:09.236503 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" event={"ID":"7986fe9f-1265-4339-ac4c-24fba0067f48","Type":"ContainerDied","Data":"8d60d7bdcd4d6b4a9e3e572101586f33cb84df92a4f495bdf1dc2e24e0f0dc95"} Oct 02 13:03:10 crc kubenswrapper[4710]: I1002 13:03:10.248073 4710 generic.go:334] "Generic (PLEG): container finished" podID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerID="8e67c8bba07c0df2c421fb329246d9c5f43628004ab6fa16f9f1189c08afa882" exitCode=0 Oct 02 13:03:10 crc kubenswrapper[4710]: I1002 13:03:10.248138 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" event={"ID":"7986fe9f-1265-4339-ac4c-24fba0067f48","Type":"ContainerDied","Data":"8e67c8bba07c0df2c421fb329246d9c5f43628004ab6fa16f9f1189c08afa882"} Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.368475 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xxrq9"] Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.368712 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xxrq9" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="registry-server" containerID="cri-o://7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e" gracePeriod=2 Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.528950 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.609910 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d755r\" (UniqueName: \"kubernetes.io/projected/7986fe9f-1265-4339-ac4c-24fba0067f48-kube-api-access-d755r\") pod \"7986fe9f-1265-4339-ac4c-24fba0067f48\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.610278 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-util\") pod \"7986fe9f-1265-4339-ac4c-24fba0067f48\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.610346 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-bundle\") pod \"7986fe9f-1265-4339-ac4c-24fba0067f48\" (UID: \"7986fe9f-1265-4339-ac4c-24fba0067f48\") " Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.611144 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-bundle" (OuterVolumeSpecName: "bundle") pod "7986fe9f-1265-4339-ac4c-24fba0067f48" (UID: "7986fe9f-1265-4339-ac4c-24fba0067f48"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.616080 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7986fe9f-1265-4339-ac4c-24fba0067f48-kube-api-access-d755r" (OuterVolumeSpecName: "kube-api-access-d755r") pod "7986fe9f-1265-4339-ac4c-24fba0067f48" (UID: "7986fe9f-1265-4339-ac4c-24fba0067f48"). InnerVolumeSpecName "kube-api-access-d755r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.620714 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-util" (OuterVolumeSpecName: "util") pod "7986fe9f-1265-4339-ac4c-24fba0067f48" (UID: "7986fe9f-1265-4339-ac4c-24fba0067f48"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.712489 4710 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.712778 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d755r\" (UniqueName: \"kubernetes.io/projected/7986fe9f-1265-4339-ac4c-24fba0067f48-kube-api-access-d755r\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:11 crc kubenswrapper[4710]: I1002 13:03:11.712855 4710 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7986fe9f-1265-4339-ac4c-24fba0067f48-util\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.264042 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" event={"ID":"7986fe9f-1265-4339-ac4c-24fba0067f48","Type":"ContainerDied","Data":"bdf8739845f5a16ab91b064b50cc88d1d526168f7d014038f8d1bf3aa864db18"} Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.264088 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdf8739845f5a16ab91b064b50cc88d1d526168f7d014038f8d1bf3aa864db18" Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.264414 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49" Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.869967 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.928779 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz5g6\" (UniqueName: \"kubernetes.io/projected/e91312f6-8834-4188-9515-f0767e8e7cd8-kube-api-access-wz5g6\") pod \"e91312f6-8834-4188-9515-f0767e8e7cd8\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.928817 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-utilities\") pod \"e91312f6-8834-4188-9515-f0767e8e7cd8\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.928867 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-catalog-content\") pod \"e91312f6-8834-4188-9515-f0767e8e7cd8\" (UID: \"e91312f6-8834-4188-9515-f0767e8e7cd8\") " Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.930030 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-utilities" (OuterVolumeSpecName: "utilities") pod "e91312f6-8834-4188-9515-f0767e8e7cd8" (UID: "e91312f6-8834-4188-9515-f0767e8e7cd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.938446 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e91312f6-8834-4188-9515-f0767e8e7cd8-kube-api-access-wz5g6" (OuterVolumeSpecName: "kube-api-access-wz5g6") pod "e91312f6-8834-4188-9515-f0767e8e7cd8" (UID: "e91312f6-8834-4188-9515-f0767e8e7cd8"). InnerVolumeSpecName "kube-api-access-wz5g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:03:12 crc kubenswrapper[4710]: I1002 13:03:12.973306 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e91312f6-8834-4188-9515-f0767e8e7cd8" (UID: "e91312f6-8834-4188-9515-f0767e8e7cd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.030627 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.030676 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91312f6-8834-4188-9515-f0767e8e7cd8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.030690 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz5g6\" (UniqueName: \"kubernetes.io/projected/e91312f6-8834-4188-9515-f0767e8e7cd8-kube-api-access-wz5g6\") on node \"crc\" DevicePath \"\"" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.272453 4710 generic.go:334] "Generic (PLEG): container finished" podID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerID="7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e" exitCode=0 Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.272502 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxrq9" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.272502 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerDied","Data":"7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e"} Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.272634 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxrq9" event={"ID":"e91312f6-8834-4188-9515-f0767e8e7cd8","Type":"ContainerDied","Data":"61b6d949e87941fd2c6997510651c29c9dcfc048e06916b792285f39202b367f"} Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.272655 4710 scope.go:117] "RemoveContainer" containerID="7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.290793 4710 scope.go:117] "RemoveContainer" containerID="76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.305434 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xxrq9"] Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.310026 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xxrq9"] Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.327803 4710 scope.go:117] "RemoveContainer" containerID="3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.346811 4710 scope.go:117] "RemoveContainer" containerID="7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e" Oct 02 13:03:13 crc kubenswrapper[4710]: E1002 13:03:13.347627 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e\": container with ID starting with 7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e not found: ID does not exist" containerID="7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.347692 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e"} err="failed to get container status \"7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e\": rpc error: code = NotFound desc = could not find container \"7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e\": container with ID starting with 7ebc7b535d7c03a8e9cac8f062ce976c1424460260c120ef7b48eb008606039e not found: ID does not exist" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.347719 4710 scope.go:117] "RemoveContainer" containerID="76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f" Oct 02 13:03:13 crc kubenswrapper[4710]: E1002 13:03:13.348115 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f\": container with ID starting with 76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f not found: ID does not exist" containerID="76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.348174 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f"} err="failed to get container status \"76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f\": rpc error: code = NotFound desc = could not find container \"76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f\": container with ID starting with 76c71df2236fdefcbb57270d7e64457f30f0b85a8ccd89347278446191dcf42f not found: ID does not exist" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.348218 4710 scope.go:117] "RemoveContainer" containerID="3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae" Oct 02 13:03:13 crc kubenswrapper[4710]: E1002 13:03:13.348686 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae\": container with ID starting with 3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae not found: ID does not exist" containerID="3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae" Oct 02 13:03:13 crc kubenswrapper[4710]: I1002 13:03:13.348712 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae"} err="failed to get container status \"3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae\": rpc error: code = NotFound desc = could not find container \"3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae\": container with ID starting with 3dd8ce11b9199e5212325ba562bfd4ec38015940b122441c446f7fd71466f0ae not found: ID does not exist" Oct 02 13:03:14 crc kubenswrapper[4710]: I1002 13:03:14.914041 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" path="/var/lib/kubelet/pods/e91312f6-8834-4188-9515-f0767e8e7cd8/volumes" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.524063 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56f696895d-565tx"] Oct 02 13:03:27 crc kubenswrapper[4710]: E1002 13:03:27.525699 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="extract" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.525727 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="extract" Oct 02 13:03:27 crc kubenswrapper[4710]: E1002 13:03:27.525767 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="util" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.525788 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="util" Oct 02 13:03:27 crc kubenswrapper[4710]: E1002 13:03:27.525805 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="registry-server" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.525813 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="registry-server" Oct 02 13:03:27 crc kubenswrapper[4710]: E1002 13:03:27.525826 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="pull" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.525834 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="pull" Oct 02 13:03:27 crc kubenswrapper[4710]: E1002 13:03:27.525849 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="extract-content" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.525858 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="extract-content" Oct 02 13:03:27 crc kubenswrapper[4710]: E1002 13:03:27.525873 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="extract-utilities" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.525882 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="extract-utilities" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.526023 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e91312f6-8834-4188-9515-f0767e8e7cd8" containerName="registry-server" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.526043 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7986fe9f-1265-4339-ac4c-24fba0067f48" containerName="extract" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.526874 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.531580 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gxgpp" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.549502 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56f696895d-565tx"] Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.642527 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjjtb\" (UniqueName: \"kubernetes.io/projected/e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c-kube-api-access-tjjtb\") pod \"openstack-operator-controller-operator-56f696895d-565tx\" (UID: \"e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c\") " pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.743708 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjjtb\" (UniqueName: \"kubernetes.io/projected/e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c-kube-api-access-tjjtb\") pod \"openstack-operator-controller-operator-56f696895d-565tx\" (UID: \"e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c\") " pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.761910 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjjtb\" (UniqueName: \"kubernetes.io/projected/e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c-kube-api-access-tjjtb\") pod \"openstack-operator-controller-operator-56f696895d-565tx\" (UID: \"e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c\") " pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:03:27 crc kubenswrapper[4710]: I1002 13:03:27.843906 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:03:28 crc kubenswrapper[4710]: I1002 13:03:28.058147 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56f696895d-565tx"] Oct 02 13:03:28 crc kubenswrapper[4710]: I1002 13:03:28.382118 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" event={"ID":"e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c","Type":"ContainerStarted","Data":"2b25bb0ebf9badf068f84f783500122c808a35d5e5c4a44c7320da267b8ec338"} Oct 02 13:03:32 crc kubenswrapper[4710]: I1002 13:03:32.409983 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" event={"ID":"e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c","Type":"ContainerStarted","Data":"5413817dd6de0e9c51e859ac514d7b2e2aac975446396849ef149fa8dc484655"} Oct 02 13:03:34 crc kubenswrapper[4710]: I1002 13:03:34.424386 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" event={"ID":"e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c","Type":"ContainerStarted","Data":"615d81d179c242d0429fe783e11468cdcab204eaa41949f269b95c2fdb7a963f"} Oct 02 13:03:34 crc kubenswrapper[4710]: I1002 13:03:34.424951 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:03:34 crc kubenswrapper[4710]: I1002 13:03:34.459196 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" podStartSLOduration=1.258398656 podStartE2EDuration="7.459174183s" podCreationTimestamp="2025-10-02 13:03:27 +0000 UTC" firstStartedPulling="2025-10-02 13:03:28.061702491 +0000 UTC m=+852.168113374" lastFinishedPulling="2025-10-02 13:03:34.262478028 +0000 UTC m=+858.368888901" observedRunningTime="2025-10-02 13:03:34.455793499 +0000 UTC m=+858.562204382" watchObservedRunningTime="2025-10-02 13:03:34.459174183 +0000 UTC m=+858.565585066" Oct 02 13:03:37 crc kubenswrapper[4710]: I1002 13:03:37.846544 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-56f696895d-565tx" Oct 02 13:04:07 crc kubenswrapper[4710]: I1002 13:04:07.984690 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s"] Oct 02 13:04:07 crc kubenswrapper[4710]: I1002 13:04:07.986613 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:07 crc kubenswrapper[4710]: I1002 13:04:07.998138 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-24lw6" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.010312 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.012232 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.018146 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-l96nq" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.024223 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.028607 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.050602 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.054152 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dfbfc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.074040 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.087802 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.093794 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-759cb"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.094805 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.100845 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gsmx7" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.107792 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-759cb"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.112211 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp8sm\" (UniqueName: \"kubernetes.io/projected/d1945e74-9d29-4e91-aab0-29ed814373bb-kube-api-access-vp8sm\") pod \"barbican-operator-controller-manager-6ff8b75857-jwd7s\" (UID: \"d1945e74-9d29-4e91-aab0-29ed814373bb\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.112285 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbj2x\" (UniqueName: \"kubernetes.io/projected/0cceb14e-b353-4c50-9790-46206743cadb-kube-api-access-lbj2x\") pod \"cinder-operator-controller-manager-644bddb6d8-4k5fd\" (UID: \"0cceb14e-b353-4c50-9790-46206743cadb\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.124936 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.125902 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.132584 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-g7jdm" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.133303 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.141202 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.145928 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.151142 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-9mrmx" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.154168 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.168804 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.170548 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.183931 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.184170 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.184976 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-zkkrs" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.185392 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.185501 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.192408 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hss54" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.214203 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdrhs\" (UniqueName: \"kubernetes.io/projected/7c8ef37b-d83a-455a-855d-64f2532c2a82-kube-api-access-gdrhs\") pod \"designate-operator-controller-manager-84f4f7b77b-nsccl\" (UID: \"7c8ef37b-d83a-455a-855d-64f2532c2a82\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.214436 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp8sm\" (UniqueName: \"kubernetes.io/projected/d1945e74-9d29-4e91-aab0-29ed814373bb-kube-api-access-vp8sm\") pod \"barbican-operator-controller-manager-6ff8b75857-jwd7s\" (UID: \"d1945e74-9d29-4e91-aab0-29ed814373bb\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.215411 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbj2x\" (UniqueName: \"kubernetes.io/projected/0cceb14e-b353-4c50-9790-46206743cadb-kube-api-access-lbj2x\") pod \"cinder-operator-controller-manager-644bddb6d8-4k5fd\" (UID: \"0cceb14e-b353-4c50-9790-46206743cadb\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.215915 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg5cc\" (UniqueName: \"kubernetes.io/projected/c7eae467-8d9a-41f5-9d91-133fffed772f-kube-api-access-gg5cc\") pod \"glance-operator-controller-manager-84958c4d49-759cb\" (UID: \"c7eae467-8d9a-41f5-9d91-133fffed772f\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.224980 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.254184 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.256099 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.260465 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.263349 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.275768 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4pjwl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.276005 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rmq6v" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.293835 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.295579 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.296672 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.298074 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp8sm\" (UniqueName: \"kubernetes.io/projected/d1945e74-9d29-4e91-aab0-29ed814373bb-kube-api-access-vp8sm\") pod \"barbican-operator-controller-manager-6ff8b75857-jwd7s\" (UID: \"d1945e74-9d29-4e91-aab0-29ed814373bb\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.298574 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbj2x\" (UniqueName: \"kubernetes.io/projected/0cceb14e-b353-4c50-9790-46206743cadb-kube-api-access-lbj2x\") pod \"cinder-operator-controller-manager-644bddb6d8-4k5fd\" (UID: \"0cceb14e-b353-4c50-9790-46206743cadb\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.311220 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-rlzj2" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318364 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm25g\" (UniqueName: \"kubernetes.io/projected/648199c6-9952-4da8-96c7-ea049a9a1d8b-kube-api-access-nm25g\") pod \"heat-operator-controller-manager-5d889d78cf-fzzx7\" (UID: \"648199c6-9952-4da8-96c7-ea049a9a1d8b\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318415 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdrhs\" (UniqueName: \"kubernetes.io/projected/7c8ef37b-d83a-455a-855d-64f2532c2a82-kube-api-access-gdrhs\") pod \"designate-operator-controller-manager-84f4f7b77b-nsccl\" (UID: \"7c8ef37b-d83a-455a-855d-64f2532c2a82\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318441 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffj6j\" (UniqueName: \"kubernetes.io/projected/44e7e977-a2d9-4cd8-8b2c-b28c62b95991-kube-api-access-ffj6j\") pod \"horizon-operator-controller-manager-9f4696d94-b96hl\" (UID: \"44e7e977-a2d9-4cd8-8b2c-b28c62b95991\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318460 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318489 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg5cc\" (UniqueName: \"kubernetes.io/projected/c7eae467-8d9a-41f5-9d91-133fffed772f-kube-api-access-gg5cc\") pod \"glance-operator-controller-manager-84958c4d49-759cb\" (UID: \"c7eae467-8d9a-41f5-9d91-133fffed772f\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318516 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrzbf\" (UniqueName: \"kubernetes.io/projected/60ef26f7-b128-4e27-a8dc-46bf65589a12-kube-api-access-xrzbf\") pod \"ironic-operator-controller-manager-5cd4858477-czjvk\" (UID: \"60ef26f7-b128-4e27-a8dc-46bf65589a12\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.318538 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvgdw\" (UniqueName: \"kubernetes.io/projected/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-kube-api-access-fvgdw\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.327838 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.329344 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.330061 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.360130 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7n6k2" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.361211 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.383354 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdrhs\" (UniqueName: \"kubernetes.io/projected/7c8ef37b-d83a-455a-855d-64f2532c2a82-kube-api-access-gdrhs\") pod \"designate-operator-controller-manager-84f4f7b77b-nsccl\" (UID: \"7c8ef37b-d83a-455a-855d-64f2532c2a82\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.383420 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.383734 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.384200 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.398080 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg5cc\" (UniqueName: \"kubernetes.io/projected/c7eae467-8d9a-41f5-9d91-133fffed772f-kube-api-access-gg5cc\") pod \"glance-operator-controller-manager-84958c4d49-759cb\" (UID: \"c7eae467-8d9a-41f5-9d91-133fffed772f\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421022 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421667 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww2xs\" (UniqueName: \"kubernetes.io/projected/fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc-kube-api-access-ww2xs\") pod \"mariadb-operator-controller-manager-88c7-zwvmz\" (UID: \"fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421736 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm25g\" (UniqueName: \"kubernetes.io/projected/648199c6-9952-4da8-96c7-ea049a9a1d8b-kube-api-access-nm25g\") pod \"heat-operator-controller-manager-5d889d78cf-fzzx7\" (UID: \"648199c6-9952-4da8-96c7-ea049a9a1d8b\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421798 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffj6j\" (UniqueName: \"kubernetes.io/projected/44e7e977-a2d9-4cd8-8b2c-b28c62b95991-kube-api-access-ffj6j\") pod \"horizon-operator-controller-manager-9f4696d94-b96hl\" (UID: \"44e7e977-a2d9-4cd8-8b2c-b28c62b95991\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421825 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421872 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghwtv\" (UniqueName: \"kubernetes.io/projected/f5b18430-0eb0-46a6-95cb-79aac8b9f170-kube-api-access-ghwtv\") pod \"neutron-operator-controller-manager-849d5b9b84-9ktdn\" (UID: \"f5b18430-0eb0-46a6-95cb-79aac8b9f170\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421905 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrzbf\" (UniqueName: \"kubernetes.io/projected/60ef26f7-b128-4e27-a8dc-46bf65589a12-kube-api-access-xrzbf\") pod \"ironic-operator-controller-manager-5cd4858477-czjvk\" (UID: \"60ef26f7-b128-4e27-a8dc-46bf65589a12\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421936 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llvl6\" (UniqueName: \"kubernetes.io/projected/e0b5b89e-972b-4961-982d-0d7b04b3e509-kube-api-access-llvl6\") pod \"keystone-operator-controller-manager-5bd55b4bff-fmkmc\" (UID: \"e0b5b89e-972b-4961-982d-0d7b04b3e509\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.421960 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvgdw\" (UniqueName: \"kubernetes.io/projected/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-kube-api-access-fvgdw\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.422001 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns5xd\" (UniqueName: \"kubernetes.io/projected/37052f75-ada0-4aa2-ba11-295678f5a627-kube-api-access-ns5xd\") pod \"manila-operator-controller-manager-6d68dbc695-2gszd\" (UID: \"37052f75-ada0-4aa2-ba11-295678f5a627\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:08 crc kubenswrapper[4710]: E1002 13:04:08.422523 4710 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 13:04:08 crc kubenswrapper[4710]: E1002 13:04:08.422571 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-cert podName:8c6c2206-348e-4fc0-8b76-4ab14e6bccb4 nodeName:}" failed. No retries permitted until 2025-10-02 13:04:08.922553961 +0000 UTC m=+893.028964844 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-cert") pod "infra-operator-controller-manager-9d6c5db85-glgh4" (UID: "8c6c2206-348e-4fc0-8b76-4ab14e6bccb4") : secret "infra-operator-webhook-server-cert" not found Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.425535 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.426912 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.433662 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.472570 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-qrmh8" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.484445 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.486615 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.492190 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-nrjxw" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.524829 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghwtv\" (UniqueName: \"kubernetes.io/projected/f5b18430-0eb0-46a6-95cb-79aac8b9f170-kube-api-access-ghwtv\") pod \"neutron-operator-controller-manager-849d5b9b84-9ktdn\" (UID: \"f5b18430-0eb0-46a6-95cb-79aac8b9f170\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.524889 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llvl6\" (UniqueName: \"kubernetes.io/projected/e0b5b89e-972b-4961-982d-0d7b04b3e509-kube-api-access-llvl6\") pod \"keystone-operator-controller-manager-5bd55b4bff-fmkmc\" (UID: \"e0b5b89e-972b-4961-982d-0d7b04b3e509\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.524932 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns5xd\" (UniqueName: \"kubernetes.io/projected/37052f75-ada0-4aa2-ba11-295678f5a627-kube-api-access-ns5xd\") pod \"manila-operator-controller-manager-6d68dbc695-2gszd\" (UID: \"37052f75-ada0-4aa2-ba11-295678f5a627\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.524958 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww2xs\" (UniqueName: \"kubernetes.io/projected/fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc-kube-api-access-ww2xs\") pod \"mariadb-operator-controller-manager-88c7-zwvmz\" (UID: \"fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.524992 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8cl6\" (UniqueName: \"kubernetes.io/projected/0d644973-169e-4c16-85a1-62680612b413-kube-api-access-m8cl6\") pod \"nova-operator-controller-manager-64cd67b5cb-mbcgh\" (UID: \"0d644973-169e-4c16-85a1-62680612b413\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.566434 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns5xd\" (UniqueName: \"kubernetes.io/projected/37052f75-ada0-4aa2-ba11-295678f5a627-kube-api-access-ns5xd\") pod \"manila-operator-controller-manager-6d68dbc695-2gszd\" (UID: \"37052f75-ada0-4aa2-ba11-295678f5a627\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.573285 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghwtv\" (UniqueName: \"kubernetes.io/projected/f5b18430-0eb0-46a6-95cb-79aac8b9f170-kube-api-access-ghwtv\") pod \"neutron-operator-controller-manager-849d5b9b84-9ktdn\" (UID: \"f5b18430-0eb0-46a6-95cb-79aac8b9f170\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.588525 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww2xs\" (UniqueName: \"kubernetes.io/projected/fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc-kube-api-access-ww2xs\") pod \"mariadb-operator-controller-manager-88c7-zwvmz\" (UID: \"fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.599640 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.630492 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7fjt\" (UniqueName: \"kubernetes.io/projected/753c96fc-cfc1-46ff-bd18-f7a41a9b8974-kube-api-access-w7fjt\") pod \"octavia-operator-controller-manager-7b787867f4-s2pcr\" (UID: \"753c96fc-cfc1-46ff-bd18-f7a41a9b8974\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.633471 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrzbf\" (UniqueName: \"kubernetes.io/projected/60ef26f7-b128-4e27-a8dc-46bf65589a12-kube-api-access-xrzbf\") pod \"ironic-operator-controller-manager-5cd4858477-czjvk\" (UID: \"60ef26f7-b128-4e27-a8dc-46bf65589a12\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.639324 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm25g\" (UniqueName: \"kubernetes.io/projected/648199c6-9952-4da8-96c7-ea049a9a1d8b-kube-api-access-nm25g\") pod \"heat-operator-controller-manager-5d889d78cf-fzzx7\" (UID: \"648199c6-9952-4da8-96c7-ea049a9a1d8b\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.639908 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvgdw\" (UniqueName: \"kubernetes.io/projected/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-kube-api-access-fvgdw\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.647655 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffj6j\" (UniqueName: \"kubernetes.io/projected/44e7e977-a2d9-4cd8-8b2c-b28c62b95991-kube-api-access-ffj6j\") pod \"horizon-operator-controller-manager-9f4696d94-b96hl\" (UID: \"44e7e977-a2d9-4cd8-8b2c-b28c62b95991\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.654022 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llvl6\" (UniqueName: \"kubernetes.io/projected/e0b5b89e-972b-4961-982d-0d7b04b3e509-kube-api-access-llvl6\") pod \"keystone-operator-controller-manager-5bd55b4bff-fmkmc\" (UID: \"e0b5b89e-972b-4961-982d-0d7b04b3e509\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.670649 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.671401 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8cl6\" (UniqueName: \"kubernetes.io/projected/0d644973-169e-4c16-85a1-62680612b413-kube-api-access-m8cl6\") pod \"nova-operator-controller-manager-64cd67b5cb-mbcgh\" (UID: \"0d644973-169e-4c16-85a1-62680612b413\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.677368 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.726597 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8cl6\" (UniqueName: \"kubernetes.io/projected/0d644973-169e-4c16-85a1-62680612b413-kube-api-access-m8cl6\") pod \"nova-operator-controller-manager-64cd67b5cb-mbcgh\" (UID: \"0d644973-169e-4c16-85a1-62680612b413\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.736336 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.737566 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.749609 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.749663 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-jd826" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.755324 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.756464 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.761781 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-hn5lh" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.772218 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7fjt\" (UniqueName: \"kubernetes.io/projected/753c96fc-cfc1-46ff-bd18-f7a41a9b8974-kube-api-access-w7fjt\") pod \"octavia-operator-controller-manager-7b787867f4-s2pcr\" (UID: \"753c96fc-cfc1-46ff-bd18-f7a41a9b8974\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.780468 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.785178 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.787492 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.788728 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.794306 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-m5chn" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.806553 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.806819 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.807040 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.816223 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.820615 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7fjt\" (UniqueName: \"kubernetes.io/projected/753c96fc-cfc1-46ff-bd18-f7a41a9b8974-kube-api-access-w7fjt\") pod \"octavia-operator-controller-manager-7b787867f4-s2pcr\" (UID: \"753c96fc-cfc1-46ff-bd18-f7a41a9b8974\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.827427 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.867124 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.878963 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x76x\" (UniqueName: \"kubernetes.io/projected/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-kube-api-access-4x76x\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.878991 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8c96\" (UniqueName: \"kubernetes.io/projected/d02bda21-fd1d-4058-9e76-7c8df8a864d3-kube-api-access-b8c96\") pod \"placement-operator-controller-manager-589c58c6c-6bwxt\" (UID: \"d02bda21-fd1d-4058-9e76-7c8df8a864d3\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.879084 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.879131 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n984z\" (UniqueName: \"kubernetes.io/projected/8a6436d8-ce16-4549-a0b5-811f677dfd0f-kube-api-access-n984z\") pod \"ovn-operator-controller-manager-9976ff44c-rh965\" (UID: \"8a6436d8-ce16-4549-a0b5-811f677dfd0f\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.887793 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.889091 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.901437 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-fh7fc" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.918350 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-nw8rd"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.920463 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.927030 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.932513 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-9bqwb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.934760 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.938576 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.939786 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.955607 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-9lbwt" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983048 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983134 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whgrw\" (UniqueName: \"kubernetes.io/projected/b360de58-b6b9-4e33-810d-574a5c0d1b70-kube-api-access-whgrw\") pod \"test-operator-controller-manager-85777745bb-nw8rd\" (UID: \"b360de58-b6b9-4e33-810d-574a5c0d1b70\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983188 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n984z\" (UniqueName: \"kubernetes.io/projected/8a6436d8-ce16-4549-a0b5-811f677dfd0f-kube-api-access-n984z\") pod \"ovn-operator-controller-manager-9976ff44c-rh965\" (UID: \"8a6436d8-ce16-4549-a0b5-811f677dfd0f\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983226 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x76x\" (UniqueName: \"kubernetes.io/projected/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-kube-api-access-4x76x\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983251 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8c96\" (UniqueName: \"kubernetes.io/projected/d02bda21-fd1d-4058-9e76-7c8df8a864d3-kube-api-access-b8c96\") pod \"placement-operator-controller-manager-589c58c6c-6bwxt\" (UID: \"d02bda21-fd1d-4058-9e76-7c8df8a864d3\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983297 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983325 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zg5s\" (UniqueName: \"kubernetes.io/projected/624fe0ee-ec13-452d-94bd-b883419e021e-kube-api-access-9zg5s\") pod \"telemetry-operator-controller-manager-b8d54b5d7-z7mcb\" (UID: \"624fe0ee-ec13-452d-94bd-b883419e021e\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.983411 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhxkj\" (UniqueName: \"kubernetes.io/projected/b4183159-214e-472b-a527-e35c091cd5bf-kube-api-access-jhxkj\") pod \"swift-operator-controller-manager-84d6b4b759-q8lfk\" (UID: \"b4183159-214e-472b-a527-e35c091cd5bf\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:08 crc kubenswrapper[4710]: E1002 13:04:08.983967 4710 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 13:04:08 crc kubenswrapper[4710]: E1002 13:04:08.984040 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-cert podName:ccccaf1d-ffb3-471e-bbc9-2f052164b7be nodeName:}" failed. No retries permitted until 2025-10-02 13:04:09.484021193 +0000 UTC m=+893.590432076 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-x4scm" (UID: "ccccaf1d-ffb3-471e-bbc9-2f052164b7be") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.988589 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk"] Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.989298 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c6c2206-348e-4fc0-8b76-4ab14e6bccb4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-glgh4\" (UID: \"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:08 crc kubenswrapper[4710]: I1002 13:04:08.993420 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-nw8rd"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:08.999931 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.001397 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.007373 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-llz98" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.008017 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8c96\" (UniqueName: \"kubernetes.io/projected/d02bda21-fd1d-4058-9e76-7c8df8a864d3-kube-api-access-b8c96\") pod \"placement-operator-controller-manager-589c58c6c-6bwxt\" (UID: \"d02bda21-fd1d-4058-9e76-7c8df8a864d3\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.010355 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.010872 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n984z\" (UniqueName: \"kubernetes.io/projected/8a6436d8-ce16-4549-a0b5-811f677dfd0f-kube-api-access-n984z\") pod \"ovn-operator-controller-manager-9976ff44c-rh965\" (UID: \"8a6436d8-ce16-4549-a0b5-811f677dfd0f\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.018378 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x76x\" (UniqueName: \"kubernetes.io/projected/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-kube-api-access-4x76x\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.024778 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.041065 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.042440 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.045097 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.045287 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-bqqkc" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.047611 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.048388 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.056086 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.057150 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.059693 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-dpm9n" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.073715 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.089263 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zg5s\" (UniqueName: \"kubernetes.io/projected/624fe0ee-ec13-452d-94bd-b883419e021e-kube-api-access-9zg5s\") pod \"telemetry-operator-controller-manager-b8d54b5d7-z7mcb\" (UID: \"624fe0ee-ec13-452d-94bd-b883419e021e\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.089305 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vjh\" (UniqueName: \"kubernetes.io/projected/7fb73401-46c9-4b5b-b6f9-c7359e67c668-kube-api-access-78vjh\") pod \"watcher-operator-controller-manager-5d597476bd-88c2w\" (UID: \"7fb73401-46c9-4b5b-b6f9-c7359e67c668\") " pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.089442 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhxkj\" (UniqueName: \"kubernetes.io/projected/b4183159-214e-472b-a527-e35c091cd5bf-kube-api-access-jhxkj\") pod \"swift-operator-controller-manager-84d6b4b759-q8lfk\" (UID: \"b4183159-214e-472b-a527-e35c091cd5bf\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.089541 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whgrw\" (UniqueName: \"kubernetes.io/projected/b360de58-b6b9-4e33-810d-574a5c0d1b70-kube-api-access-whgrw\") pod \"test-operator-controller-manager-85777745bb-nw8rd\" (UID: \"b360de58-b6b9-4e33-810d-574a5c0d1b70\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.114048 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whgrw\" (UniqueName: \"kubernetes.io/projected/b360de58-b6b9-4e33-810d-574a5c0d1b70-kube-api-access-whgrw\") pod \"test-operator-controller-manager-85777745bb-nw8rd\" (UID: \"b360de58-b6b9-4e33-810d-574a5c0d1b70\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.119145 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zg5s\" (UniqueName: \"kubernetes.io/projected/624fe0ee-ec13-452d-94bd-b883419e021e-kube-api-access-9zg5s\") pod \"telemetry-operator-controller-manager-b8d54b5d7-z7mcb\" (UID: \"624fe0ee-ec13-452d-94bd-b883419e021e\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.119274 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.120312 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhxkj\" (UniqueName: \"kubernetes.io/projected/b4183159-214e-472b-a527-e35c091cd5bf-kube-api-access-jhxkj\") pod \"swift-operator-controller-manager-84d6b4b759-q8lfk\" (UID: \"b4183159-214e-472b-a527-e35c091cd5bf\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.129380 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.143002 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.194699 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42570ebd-0936-455a-aad9-e8e37ebbe05c-cert\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.194788 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpcpv\" (UniqueName: \"kubernetes.io/projected/90363657-f227-44f6-bc20-213a4b7ab60d-kube-api-access-hpcpv\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk\" (UID: \"90363657-f227-44f6-bc20-213a4b7ab60d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.194843 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5776q\" (UniqueName: \"kubernetes.io/projected/42570ebd-0936-455a-aad9-e8e37ebbe05c-kube-api-access-5776q\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.194941 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vjh\" (UniqueName: \"kubernetes.io/projected/7fb73401-46c9-4b5b-b6f9-c7359e67c668-kube-api-access-78vjh\") pod \"watcher-operator-controller-manager-5d597476bd-88c2w\" (UID: \"7fb73401-46c9-4b5b-b6f9-c7359e67c668\") " pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.228000 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vjh\" (UniqueName: \"kubernetes.io/projected/7fb73401-46c9-4b5b-b6f9-c7359e67c668-kube-api-access-78vjh\") pod \"watcher-operator-controller-manager-5d597476bd-88c2w\" (UID: \"7fb73401-46c9-4b5b-b6f9-c7359e67c668\") " pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.252305 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.264671 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.269785 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.310096 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5776q\" (UniqueName: \"kubernetes.io/projected/42570ebd-0936-455a-aad9-e8e37ebbe05c-kube-api-access-5776q\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.310740 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42570ebd-0936-455a-aad9-e8e37ebbe05c-cert\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.310806 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpcpv\" (UniqueName: \"kubernetes.io/projected/90363657-f227-44f6-bc20-213a4b7ab60d-kube-api-access-hpcpv\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk\" (UID: \"90363657-f227-44f6-bc20-213a4b7ab60d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" Oct 02 13:04:09 crc kubenswrapper[4710]: E1002 13:04:09.311231 4710 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 13:04:09 crc kubenswrapper[4710]: E1002 13:04:09.311286 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42570ebd-0936-455a-aad9-e8e37ebbe05c-cert podName:42570ebd-0936-455a-aad9-e8e37ebbe05c nodeName:}" failed. No retries permitted until 2025-10-02 13:04:09.811268323 +0000 UTC m=+893.917679206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42570ebd-0936-455a-aad9-e8e37ebbe05c-cert") pod "openstack-operator-controller-manager-95f458bf5-dqjqf" (UID: "42570ebd-0936-455a-aad9-e8e37ebbe05c") : secret "webhook-server-cert" not found Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.328699 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.331588 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpcpv\" (UniqueName: \"kubernetes.io/projected/90363657-f227-44f6-bc20-213a4b7ab60d-kube-api-access-hpcpv\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk\" (UID: \"90363657-f227-44f6-bc20-213a4b7ab60d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.332326 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5776q\" (UniqueName: \"kubernetes.io/projected/42570ebd-0936-455a-aad9-e8e37ebbe05c-kube-api-access-5776q\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.347511 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:09 crc kubenswrapper[4710]: W1002 13:04:09.370501 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cceb14e_b353_4c50_9790_46206743cadb.slice/crio-f4de3cf022427f46452087218272cdfbc115093f517030de101b66472a077f62 WatchSource:0}: Error finding container f4de3cf022427f46452087218272cdfbc115093f517030de101b66472a077f62: Status 404 returned error can't find the container with id f4de3cf022427f46452087218272cdfbc115093f517030de101b66472a077f62 Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.382406 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.398154 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.412096 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.464428 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.513132 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.520738 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ccccaf1d-ffb3-471e-bbc9-2f052164b7be-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-x4scm\" (UID: \"ccccaf1d-ffb3-471e-bbc9-2f052164b7be\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:09 crc kubenswrapper[4710]: W1002 13:04:09.552630 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1945e74_9d29_4e91_aab0_29ed814373bb.slice/crio-8983741e82f0311363ed450e7ee1c1904501e71a5ebaacf6fc939991bb59f093 WatchSource:0}: Error finding container 8983741e82f0311363ed450e7ee1c1904501e71a5ebaacf6fc939991bb59f093: Status 404 returned error can't find the container with id 8983741e82f0311363ed450e7ee1c1904501e71a5ebaacf6fc939991bb59f093 Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.579013 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn"] Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.594674 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-759cb"] Oct 02 13:04:09 crc kubenswrapper[4710]: W1002 13:04:09.624308 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5b18430_0eb0_46a6_95cb_79aac8b9f170.slice/crio-53ecd748126ddc3eae939b33823fe7470969b329d0565f48c9225e8589ca2622 WatchSource:0}: Error finding container 53ecd748126ddc3eae939b33823fe7470969b329d0565f48c9225e8589ca2622: Status 404 returned error can't find the container with id 53ecd748126ddc3eae939b33823fe7470969b329d0565f48c9225e8589ca2622 Oct 02 13:04:09 crc kubenswrapper[4710]: W1002 13:04:09.651059 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7eae467_8d9a_41f5_9d91_133fffed772f.slice/crio-18a880f07ae3aa60dee82e3cf6a4085484db62aa22216d28f6e69e893e147105 WatchSource:0}: Error finding container 18a880f07ae3aa60dee82e3cf6a4085484db62aa22216d28f6e69e893e147105: Status 404 returned error can't find the container with id 18a880f07ae3aa60dee82e3cf6a4085484db62aa22216d28f6e69e893e147105 Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.680444 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.777461 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" event={"ID":"0cceb14e-b353-4c50-9790-46206743cadb","Type":"ContainerStarted","Data":"f4de3cf022427f46452087218272cdfbc115093f517030de101b66472a077f62"} Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.791589 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" event={"ID":"7c8ef37b-d83a-455a-855d-64f2532c2a82","Type":"ContainerStarted","Data":"31b35c4832bbd1cfe7f9eae34b60fc915527dbc3fd4acb307f3d7b1c781496f3"} Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.792593 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" event={"ID":"f5b18430-0eb0-46a6-95cb-79aac8b9f170","Type":"ContainerStarted","Data":"53ecd748126ddc3eae939b33823fe7470969b329d0565f48c9225e8589ca2622"} Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.793356 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" event={"ID":"d1945e74-9d29-4e91-aab0-29ed814373bb","Type":"ContainerStarted","Data":"8983741e82f0311363ed450e7ee1c1904501e71a5ebaacf6fc939991bb59f093"} Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.796016 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" event={"ID":"c7eae467-8d9a-41f5-9d91-133fffed772f","Type":"ContainerStarted","Data":"18a880f07ae3aa60dee82e3cf6a4085484db62aa22216d28f6e69e893e147105"} Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.820268 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42570ebd-0936-455a-aad9-e8e37ebbe05c-cert\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.825339 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42570ebd-0936-455a-aad9-e8e37ebbe05c-cert\") pod \"openstack-operator-controller-manager-95f458bf5-dqjqf\" (UID: \"42570ebd-0936-455a-aad9-e8e37ebbe05c\") " pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:09 crc kubenswrapper[4710]: I1002 13:04:09.960096 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.104006 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.138830 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.200126 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.200188 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.211400 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl"] Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.222944 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44e7e977_a2d9_4cd8_8b2c_b28c62b95991.slice/crio-8e3396f5810c65997b8faeb3786ed067ae94ac0f70aea1763ba5e1614b532a91 WatchSource:0}: Error finding container 8e3396f5810c65997b8faeb3786ed067ae94ac0f70aea1763ba5e1614b532a91: Status 404 returned error can't find the container with id 8e3396f5810c65997b8faeb3786ed067ae94ac0f70aea1763ba5e1614b532a91 Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.397103 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.404786 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.409174 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.451513 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.665047 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.674119 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.678402 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk"] Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.686706 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a6436d8_ce16_4549_a0b5_811f677dfd0f.slice/crio-efae06a6e7af7798a60a63bd6edfb3b42ae7695be174dc41041d6cee4ddf0f94 WatchSource:0}: Error finding container efae06a6e7af7798a60a63bd6edfb3b42ae7695be174dc41041d6cee4ddf0f94: Status 404 returned error can't find the container with id efae06a6e7af7798a60a63bd6edfb3b42ae7695be174dc41041d6cee4ddf0f94 Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.697954 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-nw8rd"] Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.698889 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod624fe0ee_ec13_452d_94bd_b883419e021e.slice/crio-e15a428a65f076bb62b1f5020cf7c656d9ef13cbbe0491246a03fb59430f47ef WatchSource:0}: Error finding container e15a428a65f076bb62b1f5020cf7c656d9ef13cbbe0491246a03fb59430f47ef: Status 404 returned error can't find the container with id e15a428a65f076bb62b1f5020cf7c656d9ef13cbbe0491246a03fb59430f47ef Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.699165 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90363657_f227_44f6_bc20_213a4b7ab60d.slice/crio-d4c75df4b05b74e94dcd403484d031201f42ecc9aa052eb49a87c87c85588c69 WatchSource:0}: Error finding container d4c75df4b05b74e94dcd403484d031201f42ecc9aa052eb49a87c87c85588c69: Status 404 returned error can't find the container with id d4c75df4b05b74e94dcd403484d031201f42ecc9aa052eb49a87c87c85588c69 Oct 02 13:04:10 crc kubenswrapper[4710]: E1002 13:04:10.708703 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9zg5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-z7mcb_openstack-operators(624fe0ee-ec13-452d-94bd-b883419e021e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.710500 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk"] Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.718688 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb360de58_b6b9_4e33_810d_574a5c0d1b70.slice/crio-40342c00e37a50f3e934e66acbc0a36426b9e0f7a64e0c5df9eaeb4b7f96e42d WatchSource:0}: Error finding container 40342c00e37a50f3e934e66acbc0a36426b9e0f7a64e0c5df9eaeb4b7f96e42d: Status 404 returned error can't find the container with id 40342c00e37a50f3e934e66acbc0a36426b9e0f7a64e0c5df9eaeb4b7f96e42d Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.722550 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w"] Oct 02 13:04:10 crc kubenswrapper[4710]: E1002 13:04:10.725158 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whgrw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-nw8rd_openstack-operators(b360de58-b6b9-4e33-810d-574a5c0d1b70): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.730252 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt"] Oct 02 13:04:10 crc kubenswrapper[4710]: E1002 13:04:10.743848 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b8c96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-6bwxt_openstack-operators(d02bda21-fd1d-4058-9e76-7c8df8a864d3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.749781 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fb73401_46c9_4b5b_b6f9_c7359e67c668.slice/crio-14987c60908dca262afbab2df6d9806e2e7fb8b7a3530fb17e227ab0d17f91fe WatchSource:0}: Error finding container 14987c60908dca262afbab2df6d9806e2e7fb8b7a3530fb17e227ab0d17f91fe: Status 404 returned error can't find the container with id 14987c60908dca262afbab2df6d9806e2e7fb8b7a3530fb17e227ab0d17f91fe Oct 02 13:04:10 crc kubenswrapper[4710]: E1002 13:04:10.758991 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.147:5001/openstack-k8s-operators/watcher-operator:f4db4918d3d9d309fe468fe457835ad03c41d70c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-78vjh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5d597476bd-88c2w_openstack-operators(7fb73401-46c9-4b5b-b6f9-c7359e67c668): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.764139 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4183159_214e_472b_a527_e35c091cd5bf.slice/crio-f5e7e7678ac38fc100a8c222bc5c71daee512b49ba1f6c6fd7d5a35f54392c2c WatchSource:0}: Error finding container f5e7e7678ac38fc100a8c222bc5c71daee512b49ba1f6c6fd7d5a35f54392c2c: Status 404 returned error can't find the container with id f5e7e7678ac38fc100a8c222bc5c71daee512b49ba1f6c6fd7d5a35f54392c2c Oct 02 13:04:10 crc kubenswrapper[4710]: E1002 13:04:10.767810 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jhxkj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-84d6b4b759-q8lfk_openstack-operators(b4183159-214e-472b-a527-e35c091cd5bf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.786197 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm"] Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.795687 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccccaf1d_ffb3_471e_bbc9_2f052164b7be.slice/crio-0d40d80dd23cc1c0746fd47a10af1446507b49e17e3f2db65079e364c24d253b WatchSource:0}: Error finding container 0d40d80dd23cc1c0746fd47a10af1446507b49e17e3f2db65079e364c24d253b: Status 404 returned error can't find the container with id 0d40d80dd23cc1c0746fd47a10af1446507b49e17e3f2db65079e364c24d253b Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.809731 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" event={"ID":"624fe0ee-ec13-452d-94bd-b883419e021e","Type":"ContainerStarted","Data":"e15a428a65f076bb62b1f5020cf7c656d9ef13cbbe0491246a03fb59430f47ef"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.811472 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" event={"ID":"e0b5b89e-972b-4961-982d-0d7b04b3e509","Type":"ContainerStarted","Data":"d364307745d7b55252cb47e0b2d35a36b63323feab16337fd97aca183711923f"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.813925 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" event={"ID":"0d644973-169e-4c16-85a1-62680612b413","Type":"ContainerStarted","Data":"234ccfc36d79bc9e4dbeaf0ead84fd94da9c5bd344ff39e39f0fa861e90ee9f9"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.815678 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" event={"ID":"753c96fc-cfc1-46ff-bd18-f7a41a9b8974","Type":"ContainerStarted","Data":"02633f2f91ed6782759f20f1bc41d7879ecfa2a21f6bc731cec725fdd820d30c"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.817799 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" event={"ID":"ccccaf1d-ffb3-471e-bbc9-2f052164b7be","Type":"ContainerStarted","Data":"0d40d80dd23cc1c0746fd47a10af1446507b49e17e3f2db65079e364c24d253b"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.823076 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" event={"ID":"37052f75-ada0-4aa2-ba11-295678f5a627","Type":"ContainerStarted","Data":"a2976eddcd3670cdde9b8f5f79a9aafaaaf723c39873cf1626b4a6c17a5a6351"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.825895 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" event={"ID":"60ef26f7-b128-4e27-a8dc-46bf65589a12","Type":"ContainerStarted","Data":"8d3a14f8630d8e1455bda549c0a0a9e67279db3edcf8fb43d120c80d1189c703"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.828552 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" event={"ID":"d02bda21-fd1d-4058-9e76-7c8df8a864d3","Type":"ContainerStarted","Data":"f83fa83fdb9d924b22fac067f2069264ea04338a97c43826e8a5133bdc42bd97"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.829731 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" event={"ID":"44e7e977-a2d9-4cd8-8b2c-b28c62b95991","Type":"ContainerStarted","Data":"8e3396f5810c65997b8faeb3786ed067ae94ac0f70aea1763ba5e1614b532a91"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.832456 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" event={"ID":"90363657-f227-44f6-bc20-213a4b7ab60d","Type":"ContainerStarted","Data":"d4c75df4b05b74e94dcd403484d031201f42ecc9aa052eb49a87c87c85588c69"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.834124 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" event={"ID":"8a6436d8-ce16-4549-a0b5-811f677dfd0f","Type":"ContainerStarted","Data":"efae06a6e7af7798a60a63bd6edfb3b42ae7695be174dc41041d6cee4ddf0f94"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.846079 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf"] Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.864538 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" event={"ID":"fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc","Type":"ContainerStarted","Data":"403d56a33443315efcf4380147f3273fd15873cb41b91c129aec8088b60f3145"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.871441 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" event={"ID":"b4183159-214e-472b-a527-e35c091cd5bf","Type":"ContainerStarted","Data":"f5e7e7678ac38fc100a8c222bc5c71daee512b49ba1f6c6fd7d5a35f54392c2c"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.874450 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" event={"ID":"648199c6-9952-4da8-96c7-ea049a9a1d8b","Type":"ContainerStarted","Data":"3798e68f092832b56fc073200e6cf118c80a517b4236bb21e38973ee704b4876"} Oct 02 13:04:10 crc kubenswrapper[4710]: W1002 13:04:10.877700 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42570ebd_0936_455a_aad9_e8e37ebbe05c.slice/crio-ef88eabcc5864062f66b8bb5080894b3f7199723e50b83c49fcec936c0e28738 WatchSource:0}: Error finding container ef88eabcc5864062f66b8bb5080894b3f7199723e50b83c49fcec936c0e28738: Status 404 returned error can't find the container with id ef88eabcc5864062f66b8bb5080894b3f7199723e50b83c49fcec936c0e28738 Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.877778 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" event={"ID":"b360de58-b6b9-4e33-810d-574a5c0d1b70","Type":"ContainerStarted","Data":"40342c00e37a50f3e934e66acbc0a36426b9e0f7a64e0c5df9eaeb4b7f96e42d"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.879834 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" event={"ID":"7fb73401-46c9-4b5b-b6f9-c7359e67c668","Type":"ContainerStarted","Data":"14987c60908dca262afbab2df6d9806e2e7fb8b7a3530fb17e227ab0d17f91fe"} Oct 02 13:04:10 crc kubenswrapper[4710]: I1002 13:04:10.881589 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" event={"ID":"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4","Type":"ContainerStarted","Data":"0e1ab4f058959de994c13d84798e6ccb512f51a22a32b269d7eb26cb8599a8bb"} Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.040652 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" podUID="624fe0ee-ec13-452d-94bd-b883419e021e" Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.069060 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" podUID="d02bda21-fd1d-4058-9e76-7c8df8a864d3" Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.094252 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" podUID="b4183159-214e-472b-a527-e35c091cd5bf" Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.097673 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" podUID="7fb73401-46c9-4b5b-b6f9-c7359e67c668" Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.099011 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" podUID="b360de58-b6b9-4e33-810d-574a5c0d1b70" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.896526 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" event={"ID":"7fb73401-46c9-4b5b-b6f9-c7359e67c668","Type":"ContainerStarted","Data":"4a3645ca7d2c3ded9f9736f5a361ab9f96cd23dad98daacd9fb553477ce58413"} Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.898356 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/openstack-k8s-operators/watcher-operator:f4db4918d3d9d309fe468fe457835ad03c41d70c\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" podUID="7fb73401-46c9-4b5b-b6f9-c7359e67c668" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.901263 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" event={"ID":"42570ebd-0936-455a-aad9-e8e37ebbe05c","Type":"ContainerStarted","Data":"47a975825bc7d6095fd9e6d2d708b28c4188253bfd34098b6f6c5a27c5fdaa97"} Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.901365 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" event={"ID":"42570ebd-0936-455a-aad9-e8e37ebbe05c","Type":"ContainerStarted","Data":"5f6f30876754258c037baa5f5304b0213affbfce4a4bac9a159b4c2e837dd7ab"} Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.901378 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" event={"ID":"42570ebd-0936-455a-aad9-e8e37ebbe05c","Type":"ContainerStarted","Data":"ef88eabcc5864062f66b8bb5080894b3f7199723e50b83c49fcec936c0e28738"} Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.901480 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.906881 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" event={"ID":"624fe0ee-ec13-452d-94bd-b883419e021e","Type":"ContainerStarted","Data":"8cfcc8509ba2a217658186b992311627400e64ea1dbc302695831a588d00cc06"} Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.911255 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" podUID="624fe0ee-ec13-452d-94bd-b883419e021e" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.912159 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" event={"ID":"b4183159-214e-472b-a527-e35c091cd5bf","Type":"ContainerStarted","Data":"dc230afeffb97daafb750e13b2d0fe70f586e7e11a13f3fec16d5d9b485f761e"} Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.913972 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" podUID="b4183159-214e-472b-a527-e35c091cd5bf" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.918015 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" event={"ID":"b360de58-b6b9-4e33-810d-574a5c0d1b70","Type":"ContainerStarted","Data":"4817a4d79d3dc82c443ad37c1c78c53213c839c044cb9b05675abd84ad5b433d"} Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.919631 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" podUID="b360de58-b6b9-4e33-810d-574a5c0d1b70" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.922978 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" event={"ID":"d02bda21-fd1d-4058-9e76-7c8df8a864d3","Type":"ContainerStarted","Data":"8c12ed9190688c8ae342f750a9ed83304fd6ae8051f487584e0b0859ba8ec6c3"} Oct 02 13:04:11 crc kubenswrapper[4710]: E1002 13:04:11.927175 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" podUID="d02bda21-fd1d-4058-9e76-7c8df8a864d3" Oct 02 13:04:11 crc kubenswrapper[4710]: I1002 13:04:11.980768 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" podStartSLOduration=3.980729741 podStartE2EDuration="3.980729741s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:04:11.975243154 +0000 UTC m=+896.081654037" watchObservedRunningTime="2025-10-02 13:04:11.980729741 +0000 UTC m=+896.087140624" Oct 02 13:04:12 crc kubenswrapper[4710]: E1002 13:04:12.933584 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" podUID="624fe0ee-ec13-452d-94bd-b883419e021e" Oct 02 13:04:12 crc kubenswrapper[4710]: E1002 13:04:12.933599 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" podUID="b4183159-214e-472b-a527-e35c091cd5bf" Oct 02 13:04:12 crc kubenswrapper[4710]: E1002 13:04:12.934080 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" podUID="b360de58-b6b9-4e33-810d-574a5c0d1b70" Oct 02 13:04:12 crc kubenswrapper[4710]: E1002 13:04:12.935237 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/openstack-k8s-operators/watcher-operator:f4db4918d3d9d309fe468fe457835ad03c41d70c\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" podUID="7fb73401-46c9-4b5b-b6f9-c7359e67c668" Oct 02 13:04:12 crc kubenswrapper[4710]: E1002 13:04:12.952891 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" podUID="d02bda21-fd1d-4058-9e76-7c8df8a864d3" Oct 02 13:04:19 crc kubenswrapper[4710]: I1002 13:04:19.970571 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-95f458bf5-dqjqf" Oct 02 13:04:22 crc kubenswrapper[4710]: E1002 13:04:22.904038 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884" Oct 02 13:04:22 crc kubenswrapper[4710]: E1002 13:04:22.904734 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ns5xd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-2gszd_openstack-operators(37052f75-ada0-4aa2-ba11-295678f5a627): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:04:23 crc kubenswrapper[4710]: E1002 13:04:23.409390 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a" Oct 02 13:04:23 crc kubenswrapper[4710]: E1002 13:04:23.409548 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-llvl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-5bd55b4bff-fmkmc_openstack-operators(e0b5b89e-972b-4961-982d-0d7b04b3e509): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:04:30 crc kubenswrapper[4710]: E1002 13:04:30.336812 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f" Oct 02 13:04:30 crc kubenswrapper[4710]: E1002 13:04:30.337579 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m8cl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-mbcgh_openstack-operators(0d644973-169e-4c16-85a1-62680612b413): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:04:30 crc kubenswrapper[4710]: E1002 13:04:30.389528 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8" Oct 02 13:04:30 crc kubenswrapper[4710]: E1002 13:04:30.390059 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4x76x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-x4scm_openstack-operators(ccccaf1d-ffb3-471e-bbc9-2f052164b7be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:04:30 crc kubenswrapper[4710]: E1002 13:04:30.947836 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9" Oct 02 13:04:30 crc kubenswrapper[4710]: E1002 13:04:30.948297 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w7fjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7b787867f4-s2pcr_openstack-operators(753c96fc-cfc1-46ff-bd18-f7a41a9b8974): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:04:31 crc kubenswrapper[4710]: E1002 13:04:31.311999 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" podUID="37052f75-ada0-4aa2-ba11-295678f5a627" Oct 02 13:04:31 crc kubenswrapper[4710]: E1002 13:04:31.322149 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" podUID="0d644973-169e-4c16-85a1-62680612b413" Oct 02 13:04:31 crc kubenswrapper[4710]: E1002 13:04:31.349037 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" podUID="ccccaf1d-ffb3-471e-bbc9-2f052164b7be" Oct 02 13:04:31 crc kubenswrapper[4710]: E1002 13:04:31.349154 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" podUID="753c96fc-cfc1-46ff-bd18-f7a41a9b8974" Oct 02 13:04:31 crc kubenswrapper[4710]: E1002 13:04:31.350966 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" podUID="e0b5b89e-972b-4961-982d-0d7b04b3e509" Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.074135 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" event={"ID":"e0b5b89e-972b-4961-982d-0d7b04b3e509","Type":"ContainerStarted","Data":"f9d6a6dea0938987fd38208ac7e102d369e06142c8f9fd07d7154336e4aa46b6"} Oct 02 13:04:32 crc kubenswrapper[4710]: E1002 13:04:32.076837 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" podUID="e0b5b89e-972b-4961-982d-0d7b04b3e509" Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.078985 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" event={"ID":"0d644973-169e-4c16-85a1-62680612b413","Type":"ContainerStarted","Data":"3918d007c990539b55f35e8d51aa975ef2b79d7a1a34878cf4f150e334484605"} Oct 02 13:04:32 crc kubenswrapper[4710]: E1002 13:04:32.080403 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" podUID="0d644973-169e-4c16-85a1-62680612b413" Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.083445 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" event={"ID":"44e7e977-a2d9-4cd8-8b2c-b28c62b95991","Type":"ContainerStarted","Data":"9117bc4cc75ac00b0305c88796fa49c3bafb80f2dab851a1ed4baeebdc14c010"} Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.095293 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" event={"ID":"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4","Type":"ContainerStarted","Data":"c013c2cf733a6b8582f73bf046a857f456cb6b129356e55b799593f073710903"} Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.098408 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" event={"ID":"753c96fc-cfc1-46ff-bd18-f7a41a9b8974","Type":"ContainerStarted","Data":"345f7d98448a271d22e6c5ea3b940e2bb8edce4426c2c812ab856ec65422c7d8"} Oct 02 13:04:32 crc kubenswrapper[4710]: E1002 13:04:32.101959 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" podUID="753c96fc-cfc1-46ff-bd18-f7a41a9b8974" Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.109835 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" event={"ID":"ccccaf1d-ffb3-471e-bbc9-2f052164b7be","Type":"ContainerStarted","Data":"831e43ce6326fe1e551d99d47b817538b04c58f80d1423345c48ec0198f2a6d9"} Oct 02 13:04:32 crc kubenswrapper[4710]: E1002 13:04:32.111847 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" podUID="ccccaf1d-ffb3-471e-bbc9-2f052164b7be" Oct 02 13:04:32 crc kubenswrapper[4710]: I1002 13:04:32.117869 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" event={"ID":"37052f75-ada0-4aa2-ba11-295678f5a627","Type":"ContainerStarted","Data":"309da8c59efc1ad476c0403c914f3a4450480f6aa34ddc1345746be21a89ba9f"} Oct 02 13:04:32 crc kubenswrapper[4710]: E1002 13:04:32.119494 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" podUID="37052f75-ada0-4aa2-ba11-295678f5a627" Oct 02 13:04:33 crc kubenswrapper[4710]: E1002 13:04:33.124222 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" podUID="e0b5b89e-972b-4961-982d-0d7b04b3e509" Oct 02 13:04:33 crc kubenswrapper[4710]: E1002 13:04:33.124547 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" podUID="0d644973-169e-4c16-85a1-62680612b413" Oct 02 13:04:33 crc kubenswrapper[4710]: E1002 13:04:33.125924 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" podUID="ccccaf1d-ffb3-471e-bbc9-2f052164b7be" Oct 02 13:04:33 crc kubenswrapper[4710]: E1002 13:04:33.128559 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" podUID="753c96fc-cfc1-46ff-bd18-f7a41a9b8974" Oct 02 13:04:34 crc kubenswrapper[4710]: I1002 13:04:34.132197 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" event={"ID":"f5b18430-0eb0-46a6-95cb-79aac8b9f170","Type":"ContainerStarted","Data":"6fa4d30623a87910da58f7f7b7a6e147e4fe6d9c35f97300b8ae3de382913156"} Oct 02 13:04:34 crc kubenswrapper[4710]: I1002 13:04:34.133826 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" event={"ID":"fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc","Type":"ContainerStarted","Data":"6b8970f9d12c6c2b15a6a426d9211b9b84c3a474b6fecc7fea8e092315cf6a22"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.168498 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" event={"ID":"0cceb14e-b353-4c50-9790-46206743cadb","Type":"ContainerStarted","Data":"7358b3500262fb19869a0ed0a4d20a09cd12851c47053e889f5bb024803ebd51"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.180176 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" event={"ID":"d1945e74-9d29-4e91-aab0-29ed814373bb","Type":"ContainerStarted","Data":"2fee27aa8efe1933a9418864050aab7ed8d1eb83a973641035ab860025deca71"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.191842 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" event={"ID":"60ef26f7-b128-4e27-a8dc-46bf65589a12","Type":"ContainerStarted","Data":"5a81c4f3fde0eb77878d6d5106a8d29e98ff477e4867daf63141e9861682a56d"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.201730 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" event={"ID":"7c8ef37b-d83a-455a-855d-64f2532c2a82","Type":"ContainerStarted","Data":"9f8e3dd07f67ff248f0f3274ad2842a67efdc0610cc80e8fa36137d037667754"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.206872 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" event={"ID":"90363657-f227-44f6-bc20-213a4b7ab60d","Type":"ContainerStarted","Data":"483ea611f9ba3cae6cb3b9060aeab73ad6d33755d072f44c1fd5fc0ae26f56f3"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.211860 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" event={"ID":"8a6436d8-ce16-4549-a0b5-811f677dfd0f","Type":"ContainerStarted","Data":"512e7b44d2824a8d3ad08db790b2af87258bbd1d4380d50165a262c4e65d7d2d"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.215792 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" event={"ID":"648199c6-9952-4da8-96c7-ea049a9a1d8b","Type":"ContainerStarted","Data":"9c81915b0d98f0e1038cd70f53a0cd07a0ca9e33b95e40abf2faaf338c4d3bd1"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.231529 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk" podStartSLOduration=6.914569086 podStartE2EDuration="27.231510824s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.708331101 +0000 UTC m=+894.814741984" lastFinishedPulling="2025-10-02 13:04:31.025272849 +0000 UTC m=+915.131683722" observedRunningTime="2025-10-02 13:04:35.23013884 +0000 UTC m=+919.336549723" watchObservedRunningTime="2025-10-02 13:04:35.231510824 +0000 UTC m=+919.337921707" Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.241097 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" event={"ID":"c7eae467-8d9a-41f5-9d91-133fffed772f","Type":"ContainerStarted","Data":"b7b1509a27cedfe1906d2739f6fa2d2e25966fe60975a08cd58a50e9afd3f287"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.259161 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" event={"ID":"44e7e977-a2d9-4cd8-8b2c-b28c62b95991","Type":"ContainerStarted","Data":"b651d9885b5b6d6e1037b2f6f8650c218611ee7c2552214404dc4747ecb637c1"} Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.260288 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:35 crc kubenswrapper[4710]: I1002 13:04:35.282041 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" podStartSLOduration=6.484060463 podStartE2EDuration="27.282024066s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.226080745 +0000 UTC m=+894.332491618" lastFinishedPulling="2025-10-02 13:04:31.024044338 +0000 UTC m=+915.130455221" observedRunningTime="2025-10-02 13:04:35.278969199 +0000 UTC m=+919.385380082" watchObservedRunningTime="2025-10-02 13:04:35.282024066 +0000 UTC m=+919.388434949" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.267847 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" event={"ID":"7fb73401-46c9-4b5b-b6f9-c7359e67c668","Type":"ContainerStarted","Data":"20090c4543310fff1a89b565b32100cf30c7361c9b5d39edc834ef56212a433e"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.268124 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.270271 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" event={"ID":"0cceb14e-b353-4c50-9790-46206743cadb","Type":"ContainerStarted","Data":"45d13fce41187fb8d873f6b4778781278d3bb5ea44c82bd99c49b6bca4905788"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.270468 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.272927 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" event={"ID":"b4183159-214e-472b-a527-e35c091cd5bf","Type":"ContainerStarted","Data":"e7d91c7987c17d41f66349ee520688e9fdcf42415804eb53995affa69e22c857"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.273154 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.275974 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" event={"ID":"d1945e74-9d29-4e91-aab0-29ed814373bb","Type":"ContainerStarted","Data":"57f72c3a3f5badd8c4ee5b06b04d902e55d1938f6dab5c23e5e5992582955d6b"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.276079 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.278245 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" event={"ID":"60ef26f7-b128-4e27-a8dc-46bf65589a12","Type":"ContainerStarted","Data":"f6eb2113f0f37e2f9622429e82ed535d3f8dbcaadcf46760e24aa6c335eaded4"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.278370 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.280689 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" event={"ID":"d02bda21-fd1d-4058-9e76-7c8df8a864d3","Type":"ContainerStarted","Data":"5fc9046b82fb94778ce6ade6f659b87dcf64466aef26e5b18e33fcf07a29f033"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.280916 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.282622 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" event={"ID":"8c6c2206-348e-4fc0-8b76-4ab14e6bccb4","Type":"ContainerStarted","Data":"7d82be94078e2c33c9cf34a4d00f053f2aa50be3b13c3981d6e824c1684781f3"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.282884 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.285391 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" event={"ID":"648199c6-9952-4da8-96c7-ea049a9a1d8b","Type":"ContainerStarted","Data":"39bf982c91ab08366e25914fb83b427ce60d89940a7a93ecfa1b65ff5bf37320"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.287288 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" event={"ID":"fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc","Type":"ContainerStarted","Data":"158c55ed77366839cf48a3aed1371c6f16fbfa0d0f6ed7d59a53fdb9a6aeb69a"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.287461 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.287537 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.289337 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" event={"ID":"b360de58-b6b9-4e33-810d-574a5c0d1b70","Type":"ContainerStarted","Data":"df046a573061feca60f6c5077eb64c6a15ab3a2dc0d26ce3bac6c067b2058012"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.289529 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.291636 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" event={"ID":"7c8ef37b-d83a-455a-855d-64f2532c2a82","Type":"ContainerStarted","Data":"82fa17c6cc730e18d559305ad14bdf2ca29f36630de54df6e33eebd93399d475"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.291764 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.294176 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" event={"ID":"c7eae467-8d9a-41f5-9d91-133fffed772f","Type":"ContainerStarted","Data":"1e0876fd1f3052d72902cefb541a172c70b6b02ecdbd73f00b287562a6fad78c"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.294435 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.295963 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" event={"ID":"624fe0ee-ec13-452d-94bd-b883419e021e","Type":"ContainerStarted","Data":"9e9546b401b1cc401f5c17718661d42127cdb6748203b460f364c8719a1dbb4c"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.296126 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.297852 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" event={"ID":"f5b18430-0eb0-46a6-95cb-79aac8b9f170","Type":"ContainerStarted","Data":"9cf42149424567987076057fc494a9e6c7e9c208c84df64d90dde354f4391f3b"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.300255 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" event={"ID":"8a6436d8-ce16-4549-a0b5-811f677dfd0f","Type":"ContainerStarted","Data":"0005edf03c6b953701ef186edc0ca2054e14becfcf1b0d7adc1d130707dcb035"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.300489 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.302565 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" event={"ID":"37052f75-ada0-4aa2-ba11-295678f5a627","Type":"ContainerStarted","Data":"06ea615a840d07e3ae0d22e18a7bf0f799169255cc27a9dfaf49166f7cba2fd4"} Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.304974 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-b96hl" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.309020 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" podStartSLOduration=4.110987924 podStartE2EDuration="28.308998409s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.758828051 +0000 UTC m=+894.865238934" lastFinishedPulling="2025-10-02 13:04:34.956838536 +0000 UTC m=+919.063249419" observedRunningTime="2025-10-02 13:04:36.306057645 +0000 UTC m=+920.412468528" watchObservedRunningTime="2025-10-02 13:04:36.308998409 +0000 UTC m=+920.415409292" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.357199 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" podStartSLOduration=3.509830913 podStartE2EDuration="28.357179482s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.138373518 +0000 UTC m=+894.244784401" lastFinishedPulling="2025-10-02 13:04:34.985722087 +0000 UTC m=+919.092132970" observedRunningTime="2025-10-02 13:04:36.333839709 +0000 UTC m=+920.440250612" watchObservedRunningTime="2025-10-02 13:04:36.357179482 +0000 UTC m=+920.463590365" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.358013 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" podStartSLOduration=6.980533618 podStartE2EDuration="28.358009293s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:09.658482801 +0000 UTC m=+893.764893684" lastFinishedPulling="2025-10-02 13:04:31.035958476 +0000 UTC m=+915.142369359" observedRunningTime="2025-10-02 13:04:36.355233783 +0000 UTC m=+920.461644676" watchObservedRunningTime="2025-10-02 13:04:36.358009293 +0000 UTC m=+920.464420176" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.403503 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" podStartSLOduration=7.765454876 podStartE2EDuration="29.403480478s" podCreationTimestamp="2025-10-02 13:04:07 +0000 UTC" firstStartedPulling="2025-10-02 13:04:09.397733309 +0000 UTC m=+893.504144192" lastFinishedPulling="2025-10-02 13:04:31.035758911 +0000 UTC m=+915.142169794" observedRunningTime="2025-10-02 13:04:36.382360581 +0000 UTC m=+920.488771464" watchObservedRunningTime="2025-10-02 13:04:36.403480478 +0000 UTC m=+920.509891361" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.406068 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" podStartSLOduration=7.016433845 podStartE2EDuration="28.406051012s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:09.627457157 +0000 UTC m=+893.733868040" lastFinishedPulling="2025-10-02 13:04:31.017074324 +0000 UTC m=+915.123485207" observedRunningTime="2025-10-02 13:04:36.398851432 +0000 UTC m=+920.505262315" watchObservedRunningTime="2025-10-02 13:04:36.406051012 +0000 UTC m=+920.512461895" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.445201 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" podStartSLOduration=8.117214966 podStartE2EDuration="28.445178639s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.689211294 +0000 UTC m=+894.795622177" lastFinishedPulling="2025-10-02 13:04:31.017174967 +0000 UTC m=+915.123585850" observedRunningTime="2025-10-02 13:04:36.440567824 +0000 UTC m=+920.546978727" watchObservedRunningTime="2025-10-02 13:04:36.445178639 +0000 UTC m=+920.551589522" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.482007 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" podStartSLOduration=4.359264363 podStartE2EDuration="28.481973928s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.767660811 +0000 UTC m=+894.874071684" lastFinishedPulling="2025-10-02 13:04:34.890370366 +0000 UTC m=+918.996781249" observedRunningTime="2025-10-02 13:04:36.481229179 +0000 UTC m=+920.587640062" watchObservedRunningTime="2025-10-02 13:04:36.481973928 +0000 UTC m=+920.588384811" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.506369 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-glgh4" podStartSLOduration=7.992535369 podStartE2EDuration="28.506344906s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.493252487 +0000 UTC m=+894.599663380" lastFinishedPulling="2025-10-02 13:04:31.007062034 +0000 UTC m=+915.113472917" observedRunningTime="2025-10-02 13:04:36.501995318 +0000 UTC m=+920.608406211" watchObservedRunningTime="2025-10-02 13:04:36.506344906 +0000 UTC m=+920.612755799" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.525601 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" podStartSLOduration=4.359855229 podStartE2EDuration="28.525582357s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.724984377 +0000 UTC m=+894.831395270" lastFinishedPulling="2025-10-02 13:04:34.890711515 +0000 UTC m=+918.997122398" observedRunningTime="2025-10-02 13:04:36.524837698 +0000 UTC m=+920.631248591" watchObservedRunningTime="2025-10-02 13:04:36.525582357 +0000 UTC m=+920.631993250" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.591508 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" podStartSLOduration=7.786030301 podStartE2EDuration="28.591490002s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.219929421 +0000 UTC m=+894.326340314" lastFinishedPulling="2025-10-02 13:04:31.025389132 +0000 UTC m=+915.131800015" observedRunningTime="2025-10-02 13:04:36.581260097 +0000 UTC m=+920.687670990" watchObservedRunningTime="2025-10-02 13:04:36.591490002 +0000 UTC m=+920.697900885" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.610348 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" podStartSLOduration=4.460185373 podStartE2EDuration="28.610330503s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.743645462 +0000 UTC m=+894.850056345" lastFinishedPulling="2025-10-02 13:04:34.893790582 +0000 UTC m=+919.000201475" observedRunningTime="2025-10-02 13:04:36.607657796 +0000 UTC m=+920.714068709" watchObservedRunningTime="2025-10-02 13:04:36.610330503 +0000 UTC m=+920.716741386" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.664631 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" podStartSLOduration=7.846834991 podStartE2EDuration="28.664613088s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.207139163 +0000 UTC m=+894.313550046" lastFinishedPulling="2025-10-02 13:04:31.02491726 +0000 UTC m=+915.131328143" observedRunningTime="2025-10-02 13:04:36.642452925 +0000 UTC m=+920.748863818" watchObservedRunningTime="2025-10-02 13:04:36.664613088 +0000 UTC m=+920.771023971" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.666228 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" podStartSLOduration=8.209860347 podStartE2EDuration="29.666220818s" podCreationTimestamp="2025-10-02 13:04:07 +0000 UTC" firstStartedPulling="2025-10-02 13:04:09.567482802 +0000 UTC m=+893.673893685" lastFinishedPulling="2025-10-02 13:04:31.023843273 +0000 UTC m=+915.130254156" observedRunningTime="2025-10-02 13:04:36.661727466 +0000 UTC m=+920.768138349" watchObservedRunningTime="2025-10-02 13:04:36.666220818 +0000 UTC m=+920.772631701" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.685381 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" podStartSLOduration=4.500644134 podStartE2EDuration="28.685361396s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.708562987 +0000 UTC m=+894.814973870" lastFinishedPulling="2025-10-02 13:04:34.893280249 +0000 UTC m=+918.999691132" observedRunningTime="2025-10-02 13:04:36.679925841 +0000 UTC m=+920.786336724" watchObservedRunningTime="2025-10-02 13:04:36.685361396 +0000 UTC m=+920.791772289" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.704007 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" podStartSLOduration=8.157722577 podStartE2EDuration="29.703987811s" podCreationTimestamp="2025-10-02 13:04:07 +0000 UTC" firstStartedPulling="2025-10-02 13:04:09.507831325 +0000 UTC m=+893.614242208" lastFinishedPulling="2025-10-02 13:04:31.054096559 +0000 UTC m=+915.160507442" observedRunningTime="2025-10-02 13:04:36.702420022 +0000 UTC m=+920.808830915" watchObservedRunningTime="2025-10-02 13:04:36.703987811 +0000 UTC m=+920.810398694" Oct 02 13:04:36 crc kubenswrapper[4710]: I1002 13:04:36.735716 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" podStartSLOduration=8.171449587 podStartE2EDuration="28.735699313s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.460051939 +0000 UTC m=+894.566462822" lastFinishedPulling="2025-10-02 13:04:31.024301665 +0000 UTC m=+915.130712548" observedRunningTime="2025-10-02 13:04:36.73116363 +0000 UTC m=+920.837574503" watchObservedRunningTime="2025-10-02 13:04:36.735699313 +0000 UTC m=+920.842110196" Oct 02 13:04:37 crc kubenswrapper[4710]: I1002 13:04:37.321248 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:37 crc kubenswrapper[4710]: I1002 13:04:37.321630 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:38 crc kubenswrapper[4710]: I1002 13:04:38.324490 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-9ktdn" Oct 02 13:04:38 crc kubenswrapper[4710]: I1002 13:04:38.828596 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:38 crc kubenswrapper[4710]: I1002 13:04:38.872060 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-zwvmz" Oct 02 13:04:39 crc kubenswrapper[4710]: I1002 13:04:39.123359 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rh965" Oct 02 13:04:45 crc kubenswrapper[4710]: I1002 13:04:45.385406 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" event={"ID":"ccccaf1d-ffb3-471e-bbc9-2f052164b7be","Type":"ContainerStarted","Data":"b527f885cb1a8f8a876ed372e66beeb7adae4ae1301b0c439df01f6e4450ce2f"} Oct 02 13:04:45 crc kubenswrapper[4710]: I1002 13:04:45.386700 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:45 crc kubenswrapper[4710]: I1002 13:04:45.434199 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" podStartSLOduration=3.666113652 podStartE2EDuration="37.434176936s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.813589986 +0000 UTC m=+894.920000869" lastFinishedPulling="2025-10-02 13:04:44.58165326 +0000 UTC m=+928.688064153" observedRunningTime="2025-10-02 13:04:45.430262048 +0000 UTC m=+929.536672941" watchObservedRunningTime="2025-10-02 13:04:45.434176936 +0000 UTC m=+929.540587829" Oct 02 13:04:46 crc kubenswrapper[4710]: I1002 13:04:46.394018 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" event={"ID":"753c96fc-cfc1-46ff-bd18-f7a41a9b8974","Type":"ContainerStarted","Data":"fd98fd70532d278a8d9463917e925bc082ec02e13881978f23bac25f3c92d0b6"} Oct 02 13:04:46 crc kubenswrapper[4710]: I1002 13:04:46.395609 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:04:46 crc kubenswrapper[4710]: I1002 13:04:46.415524 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" podStartSLOduration=3.47413929 podStartE2EDuration="38.415502219s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.44203723 +0000 UTC m=+894.548448123" lastFinishedPulling="2025-10-02 13:04:45.383400169 +0000 UTC m=+929.489811052" observedRunningTime="2025-10-02 13:04:46.410012212 +0000 UTC m=+930.516423095" watchObservedRunningTime="2025-10-02 13:04:46.415502219 +0000 UTC m=+930.521913122" Oct 02 13:04:47 crc kubenswrapper[4710]: I1002 13:04:47.403146 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" event={"ID":"e0b5b89e-972b-4961-982d-0d7b04b3e509","Type":"ContainerStarted","Data":"72b404ef29afb9c3eb26cbb8dfe20940856b08adaf473be5b97108c43857de6b"} Oct 02 13:04:47 crc kubenswrapper[4710]: I1002 13:04:47.404212 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:47 crc kubenswrapper[4710]: I1002 13:04:47.426705 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" podStartSLOduration=3.190599169 podStartE2EDuration="39.426684688s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.21985691 +0000 UTC m=+894.326267793" lastFinishedPulling="2025-10-02 13:04:46.455942429 +0000 UTC m=+930.562353312" observedRunningTime="2025-10-02 13:04:47.424196245 +0000 UTC m=+931.530607148" watchObservedRunningTime="2025-10-02 13:04:47.426684688 +0000 UTC m=+931.533095571" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.334065 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-jwd7s" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.388166 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-4k5fd" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.390613 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-nsccl" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.442280 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-759cb" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.788783 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-fzzx7" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.830450 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2gszd" Oct 02 13:04:48 crc kubenswrapper[4710]: I1002 13:04:48.945629 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-czjvk" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.145690 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-6bwxt" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.267248 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-z7mcb" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.272278 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-nw8rd" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.332362 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-q8lfk" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.354328 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d597476bd-88c2w" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.429731 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" event={"ID":"0d644973-169e-4c16-85a1-62680612b413","Type":"ContainerStarted","Data":"b05f2094e2e992e4840720ab522aa98dcd6cf92421bae88de157125861c5c16b"} Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.429928 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.447119 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" podStartSLOduration=3.531327748 podStartE2EDuration="41.447068814s" podCreationTimestamp="2025-10-02 13:04:08 +0000 UTC" firstStartedPulling="2025-10-02 13:04:10.435802045 +0000 UTC m=+894.542212938" lastFinishedPulling="2025-10-02 13:04:48.351543081 +0000 UTC m=+932.457954004" observedRunningTime="2025-10-02 13:04:49.444715635 +0000 UTC m=+933.551126538" watchObservedRunningTime="2025-10-02 13:04:49.447068814 +0000 UTC m=+933.553479697" Oct 02 13:04:49 crc kubenswrapper[4710]: I1002 13:04:49.688046 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-x4scm" Oct 02 13:04:58 crc kubenswrapper[4710]: I1002 13:04:58.784695 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-fmkmc" Oct 02 13:04:59 crc kubenswrapper[4710]: I1002 13:04:59.027714 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mbcgh" Oct 02 13:04:59 crc kubenswrapper[4710]: I1002 13:04:59.051252 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-s2pcr" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.648957 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784b55c5d9-mzqnv"] Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.651037 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.653175 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.653301 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.653736 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.653815 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8vrnp" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.663463 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784b55c5d9-mzqnv"] Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.704369 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bf56b5889-6j2b4"] Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.706293 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.713449 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.721944 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bf56b5889-6j2b4"] Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.824177 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-config\") pod \"dnsmasq-dns-784b55c5d9-mzqnv\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.824446 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-config\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.824560 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl859\" (UniqueName: \"kubernetes.io/projected/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-kube-api-access-fl859\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.824660 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-dns-svc\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.824764 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gqhc\" (UniqueName: \"kubernetes.io/projected/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-kube-api-access-7gqhc\") pod \"dnsmasq-dns-784b55c5d9-mzqnv\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.926535 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-config\") pod \"dnsmasq-dns-784b55c5d9-mzqnv\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.926614 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-config\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.926646 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl859\" (UniqueName: \"kubernetes.io/projected/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-kube-api-access-fl859\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.926736 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-dns-svc\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.927012 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gqhc\" (UniqueName: \"kubernetes.io/projected/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-kube-api-access-7gqhc\") pod \"dnsmasq-dns-784b55c5d9-mzqnv\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.927478 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-config\") pod \"dnsmasq-dns-784b55c5d9-mzqnv\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.928139 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-config\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.942326 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-dns-svc\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.946978 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl859\" (UniqueName: \"kubernetes.io/projected/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-kube-api-access-fl859\") pod \"dnsmasq-dns-bf56b5889-6j2b4\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.961996 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gqhc\" (UniqueName: \"kubernetes.io/projected/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-kube-api-access-7gqhc\") pod \"dnsmasq-dns-784b55c5d9-mzqnv\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:18 crc kubenswrapper[4710]: I1002 13:05:18.975030 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:19 crc kubenswrapper[4710]: I1002 13:05:19.031448 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:19 crc kubenswrapper[4710]: I1002 13:05:19.453735 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784b55c5d9-mzqnv"] Oct 02 13:05:19 crc kubenswrapper[4710]: I1002 13:05:19.561514 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bf56b5889-6j2b4"] Oct 02 13:05:19 crc kubenswrapper[4710]: W1002 13:05:19.571291 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7d7b97d_c9fb_4d08_8c34_df63eefd96a7.slice/crio-65e867df1725a262bcb391379fc9111afbee5429621273d5b163c8acff61a695 WatchSource:0}: Error finding container 65e867df1725a262bcb391379fc9111afbee5429621273d5b163c8acff61a695: Status 404 returned error can't find the container with id 65e867df1725a262bcb391379fc9111afbee5429621273d5b163c8acff61a695 Oct 02 13:05:19 crc kubenswrapper[4710]: I1002 13:05:19.663624 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" event={"ID":"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6","Type":"ContainerStarted","Data":"c8c0f361ae2eaec06461213937dec82f1c3cd55b10345321d5cc625b116cbff6"} Oct 02 13:05:19 crc kubenswrapper[4710]: I1002 13:05:19.664914 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" event={"ID":"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7","Type":"ContainerStarted","Data":"65e867df1725a262bcb391379fc9111afbee5429621273d5b163c8acff61a695"} Oct 02 13:05:22 crc kubenswrapper[4710]: I1002 13:05:22.530598 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:05:22 crc kubenswrapper[4710]: I1002 13:05:22.531107 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:05:22 crc kubenswrapper[4710]: I1002 13:05:22.931189 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784b55c5d9-mzqnv"] Oct 02 13:05:22 crc kubenswrapper[4710]: I1002 13:05:22.955796 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7985db964c-92ssp"] Oct 02 13:05:22 crc kubenswrapper[4710]: I1002 13:05:22.957025 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:22 crc kubenswrapper[4710]: I1002 13:05:22.973481 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7985db964c-92ssp"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.103844 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8pth\" (UniqueName: \"kubernetes.io/projected/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-kube-api-access-s8pth\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.103920 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-config\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.103951 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-dns-svc\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.205017 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8pth\" (UniqueName: \"kubernetes.io/projected/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-kube-api-access-s8pth\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.205113 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-config\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.205152 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-dns-svc\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.206108 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-config\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.206142 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-dns-svc\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.227406 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8pth\" (UniqueName: \"kubernetes.io/projected/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-kube-api-access-s8pth\") pod \"dnsmasq-dns-7985db964c-92ssp\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.254734 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bf56b5889-6j2b4"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.275290 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.281547 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6b645ddc-l5f42"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.283078 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.291987 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6b645ddc-l5f42"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.307152 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-config\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.307206 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5lnd\" (UniqueName: \"kubernetes.io/projected/5ed3b611-ed83-4cce-843e-b64b21afecb6-kube-api-access-w5lnd\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.307286 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-dns-svc\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.409538 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-dns-svc\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.409640 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-config\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.409674 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5lnd\" (UniqueName: \"kubernetes.io/projected/5ed3b611-ed83-4cce-843e-b64b21afecb6-kube-api-access-w5lnd\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.410820 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-dns-svc\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.411712 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-config\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.451913 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5lnd\" (UniqueName: \"kubernetes.io/projected/5ed3b611-ed83-4cce-843e-b64b21afecb6-kube-api-access-w5lnd\") pod \"dnsmasq-dns-b6b645ddc-l5f42\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.611115 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.632081 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7985db964c-92ssp"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.658001 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bb7dbd587-vtwc9"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.659676 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.683837 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bb7dbd587-vtwc9"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.818935 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-config\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.819080 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzx99\" (UniqueName: \"kubernetes.io/projected/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-kube-api-access-jzx99\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.819124 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-dns-svc\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.891370 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7985db964c-92ssp"] Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.920801 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-config\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.920897 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzx99\" (UniqueName: \"kubernetes.io/projected/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-kube-api-access-jzx99\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.920920 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-dns-svc\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.922123 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-dns-svc\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.922626 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-config\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:23 crc kubenswrapper[4710]: I1002 13:05:23.950669 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzx99\" (UniqueName: \"kubernetes.io/projected/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-kube-api-access-jzx99\") pod \"dnsmasq-dns-5bb7dbd587-vtwc9\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.006974 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.126360 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.142458 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.142871 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.145389 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.149362 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.156153 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.156387 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gzd98" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.156521 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.156670 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.156839 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.171434 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6b645ddc-l5f42"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326520 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7fb5713-6a35-4676-b466-6a43a4037f7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326601 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326645 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326672 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326707 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghww\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-kube-api-access-xghww\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326734 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326776 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326791 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326812 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326853 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.326869 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7fb5713-6a35-4676-b466-6a43a4037f7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.410721 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.411968 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.417091 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.417662 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-b5qhl" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.419511 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.419579 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.419682 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.419717 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.421661 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.426944 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.427510 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.428044 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.428199 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.428338 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xghww\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-kube-api-access-xghww\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.428778 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.428901 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429311 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429468 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429588 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429687 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7fb5713-6a35-4676-b466-6a43a4037f7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429847 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7fb5713-6a35-4676-b466-6a43a4037f7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.430132 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429357 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.431345 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429108 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.429739 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.436408 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7fb5713-6a35-4676-b466-6a43a4037f7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.443637 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.444603 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.450644 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.450929 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7fb5713-6a35-4676-b466-6a43a4037f7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.454136 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghww\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-kube-api-access-xghww\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.458671 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.469924 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531503 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531600 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531623 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531640 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531657 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531674 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlzmk\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-kube-api-access-rlzmk\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531690 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531708 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531782 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531851 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.531900 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.632971 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633033 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633062 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633081 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlzmk\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-kube-api-access-rlzmk\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633103 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633127 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633145 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633177 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633208 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633364 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.633441 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.634279 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.635566 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.635875 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.643251 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.643378 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.644386 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.647792 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.648823 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.650065 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.650341 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.661474 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlzmk\" (UniqueName: \"kubernetes.io/projected/98c2f825-d5c6-47e7-a96d-8cf7227b0a24-kube-api-access-rlzmk\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.676541 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"98c2f825-d5c6-47e7-a96d-8cf7227b0a24\") " pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.733833 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.766126 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.769038 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.775180 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.775203 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x6wq4" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.775570 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.775685 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.775928 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.775781 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.776237 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.784260 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936573 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936659 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/770af87f-9638-4cdb-a10d-2b40d50302ca-pod-info\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936693 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936772 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936798 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-config-data\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936831 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936869 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936926 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936948 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-server-conf\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.936971 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2fjs\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-kube-api-access-m2fjs\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:24 crc kubenswrapper[4710]: I1002 13:05:24.937003 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/770af87f-9638-4cdb-a10d-2b40d50302ca-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038072 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/770af87f-9638-4cdb-a10d-2b40d50302ca-pod-info\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038140 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038211 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038238 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-config-data\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038280 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038320 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038346 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038366 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-server-conf\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038387 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2fjs\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-kube-api-access-m2fjs\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038413 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/770af87f-9638-4cdb-a10d-2b40d50302ca-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038534 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.038985 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.039062 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.039617 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.040088 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-server-conf\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.040536 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-config-data\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.042013 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.043216 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.046536 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/770af87f-9638-4cdb-a10d-2b40d50302ca-pod-info\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.047316 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.048843 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/770af87f-9638-4cdb-a10d-2b40d50302ca-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.056827 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2fjs\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-kube-api-access-m2fjs\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.097189 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " pod="openstack/rabbitmq-server-0" Oct 02 13:05:25 crc kubenswrapper[4710]: I1002 13:05:25.101144 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.281120 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.282663 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.285192 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.285274 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.285382 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.285513 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-d95bx" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.285810 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.292052 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.307264 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.381934 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.382451 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-kolla-config\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.382647 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-config-data-default\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.382773 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.382874 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6lp2\" (UniqueName: \"kubernetes.io/projected/40dcfd31-21c8-4581-b2f6-889737eea66d-kube-api-access-c6lp2\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.382983 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.383082 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-secrets\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.383193 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.383301 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/40dcfd31-21c8-4581-b2f6-889737eea66d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484717 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484801 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/40dcfd31-21c8-4581-b2f6-889737eea66d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484847 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484873 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-kolla-config\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484914 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-config-data-default\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484932 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484951 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6lp2\" (UniqueName: \"kubernetes.io/projected/40dcfd31-21c8-4581-b2f6-889737eea66d-kube-api-access-c6lp2\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484969 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.484986 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-secrets\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.486145 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.486712 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-config-data-default\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.487072 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/40dcfd31-21c8-4581-b2f6-889737eea66d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.487297 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-kolla-config\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.487696 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40dcfd31-21c8-4581-b2f6-889737eea66d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.489999 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-secrets\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.496468 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.496718 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40dcfd31-21c8-4581-b2f6-889737eea66d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.505307 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6lp2\" (UniqueName: \"kubernetes.io/projected/40dcfd31-21c8-4581-b2f6-889737eea66d-kube-api-access-c6lp2\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.520295 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"40dcfd31-21c8-4581-b2f6-889737eea66d\") " pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.613833 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.678110 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.679563 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.681335 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-968rg" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.681382 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.681342 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.682432 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.702391 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.788786 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.788838 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789073 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2649\" (UniqueName: \"kubernetes.io/projected/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-kube-api-access-j2649\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789146 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789170 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789216 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789242 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789281 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.789347 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901395 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901482 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2649\" (UniqueName: \"kubernetes.io/projected/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-kube-api-access-j2649\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901517 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901538 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901581 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901621 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901691 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901794 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.901828 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.902397 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.903219 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.903460 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.903781 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.904941 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.907577 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.909335 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.911724 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.934375 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2649\" (UniqueName: \"kubernetes.io/projected/ada9e73f-75af-4b4a-8e52-af2d5b8c3728-kube-api-access-j2649\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:27 crc kubenswrapper[4710]: I1002 13:05:27.959138 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ada9e73f-75af-4b4a-8e52-af2d5b8c3728\") " pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.011424 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.083471 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.084421 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.086957 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-fx5dl" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.087122 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.087383 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.099736 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.207098 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5d102b1-889f-487b-be09-725834944ff6-config-data\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.207149 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5d102b1-889f-487b-be09-725834944ff6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.207370 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5d102b1-889f-487b-be09-725834944ff6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.207448 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c5d102b1-889f-487b-be09-725834944ff6-kolla-config\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.207483 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9tkb\" (UniqueName: \"kubernetes.io/projected/c5d102b1-889f-487b-be09-725834944ff6-kube-api-access-k9tkb\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.309170 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5d102b1-889f-487b-be09-725834944ff6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.309261 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c5d102b1-889f-487b-be09-725834944ff6-kolla-config\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.309303 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9tkb\" (UniqueName: \"kubernetes.io/projected/c5d102b1-889f-487b-be09-725834944ff6-kube-api-access-k9tkb\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.309469 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5d102b1-889f-487b-be09-725834944ff6-config-data\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.309515 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5d102b1-889f-487b-be09-725834944ff6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.310520 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c5d102b1-889f-487b-be09-725834944ff6-kolla-config\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.310651 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5d102b1-889f-487b-be09-725834944ff6-config-data\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.315119 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5d102b1-889f-487b-be09-725834944ff6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.320554 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5d102b1-889f-487b-be09-725834944ff6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.332409 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9tkb\" (UniqueName: \"kubernetes.io/projected/c5d102b1-889f-487b-be09-725834944ff6-kube-api-access-k9tkb\") pod \"memcached-0\" (UID: \"c5d102b1-889f-487b-be09-725834944ff6\") " pod="openstack/memcached-0" Oct 02 13:05:28 crc kubenswrapper[4710]: I1002 13:05:28.400483 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.751359 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.752900 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.755054 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qgsh2" Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.762690 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.856661 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54fm2\" (UniqueName: \"kubernetes.io/projected/27670d96-0e34-41ae-a54c-2bb7f19b1c6e-kube-api-access-54fm2\") pod \"kube-state-metrics-0\" (UID: \"27670d96-0e34-41ae-a54c-2bb7f19b1c6e\") " pod="openstack/kube-state-metrics-0" Oct 02 13:05:29 crc kubenswrapper[4710]: W1002 13:05:29.940597 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dfff010_9bbd_4c5d_846d_50e39fe0e12e.slice/crio-853998f71a7f3e7b61f1974e58415e2eb907f0c7b0c50df8d47aaf9714322936 WatchSource:0}: Error finding container 853998f71a7f3e7b61f1974e58415e2eb907f0c7b0c50df8d47aaf9714322936: Status 404 returned error can't find the container with id 853998f71a7f3e7b61f1974e58415e2eb907f0c7b0c50df8d47aaf9714322936 Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.958116 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54fm2\" (UniqueName: \"kubernetes.io/projected/27670d96-0e34-41ae-a54c-2bb7f19b1c6e-kube-api-access-54fm2\") pod \"kube-state-metrics-0\" (UID: \"27670d96-0e34-41ae-a54c-2bb7f19b1c6e\") " pod="openstack/kube-state-metrics-0" Oct 02 13:05:29 crc kubenswrapper[4710]: I1002 13:05:29.987908 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54fm2\" (UniqueName: \"kubernetes.io/projected/27670d96-0e34-41ae-a54c-2bb7f19b1c6e-kube-api-access-54fm2\") pod \"kube-state-metrics-0\" (UID: \"27670d96-0e34-41ae-a54c-2bb7f19b1c6e\") " pod="openstack/kube-state-metrics-0" Oct 02 13:05:30 crc kubenswrapper[4710]: I1002 13:05:30.079719 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 13:05:30 crc kubenswrapper[4710]: I1002 13:05:30.791201 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" event={"ID":"5ed3b611-ed83-4cce-843e-b64b21afecb6","Type":"ContainerStarted","Data":"b60ac86f16f030076ab4f2a50eb0269d443796bbaf9f2507c219cd90c40829a4"} Oct 02 13:05:30 crc kubenswrapper[4710]: I1002 13:05:30.797788 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7985db964c-92ssp" event={"ID":"0dfff010-9bbd-4c5d-846d-50e39fe0e12e","Type":"ContainerStarted","Data":"853998f71a7f3e7b61f1974e58415e2eb907f0c7b0c50df8d47aaf9714322936"} Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.063413 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.065423 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.067906 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.068117 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.068264 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.068421 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-58df2" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.068800 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.075548 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.081387 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180370 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8530440e-9b9d-40bc-95ac-b7c40899b0cb-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180446 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180512 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180564 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180609 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180641 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180681 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.180777 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-274w2\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-kube-api-access-274w2\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.282947 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283019 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283083 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283285 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-274w2\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-kube-api-access-274w2\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283377 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8530440e-9b9d-40bc-95ac-b7c40899b0cb-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283420 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283474 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.283530 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.284811 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8530440e-9b9d-40bc-95ac-b7c40899b0cb-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.288018 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.288115 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.290477 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.293789 4710 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.293843 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/549ec1f26b5b9c00df36a940184c18d591918f74b3d89fea9d49d0eb0309560f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.296153 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.296318 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.306989 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-274w2\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-kube-api-access-274w2\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.328635 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:31 crc kubenswrapper[4710]: I1002 13:05:31.388724 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.492137 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.494024 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.496436 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.496465 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.496539 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.497553 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.498322 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-pw78m" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.504836 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.619705 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.619778 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbac60db-a726-44d9-998b-cfca036d65d9-config\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.619922 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.620012 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.620108 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pmv7\" (UniqueName: \"kubernetes.io/projected/cbac60db-a726-44d9-998b-cfca036d65d9-kube-api-access-7pmv7\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.620240 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cbac60db-a726-44d9-998b-cfca036d65d9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.620275 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.620338 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbac60db-a726-44d9-998b-cfca036d65d9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.721996 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722056 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbac60db-a726-44d9-998b-cfca036d65d9-config\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722082 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722107 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722159 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pmv7\" (UniqueName: \"kubernetes.io/projected/cbac60db-a726-44d9-998b-cfca036d65d9-kube-api-access-7pmv7\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722194 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cbac60db-a726-44d9-998b-cfca036d65d9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722210 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722236 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbac60db-a726-44d9-998b-cfca036d65d9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722616 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.722738 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cbac60db-a726-44d9-998b-cfca036d65d9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.723653 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbac60db-a726-44d9-998b-cfca036d65d9-config\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.723844 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbac60db-a726-44d9-998b-cfca036d65d9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.727723 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.734396 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.744614 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbac60db-a726-44d9-998b-cfca036d65d9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.744934 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pmv7\" (UniqueName: \"kubernetes.io/projected/cbac60db-a726-44d9-998b-cfca036d65d9-kube-api-access-7pmv7\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.806959 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cbac60db-a726-44d9-998b-cfca036d65d9\") " pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:33 crc kubenswrapper[4710]: I1002 13:05:33.812866 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.947303 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fp7wl"] Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.949580 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.953262 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.953460 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.953552 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-whz5w" Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.965554 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jd2lt"] Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.968010 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:34 crc kubenswrapper[4710]: I1002 13:05:34.987147 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jd2lt"] Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.011338 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fp7wl"] Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050656 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-run\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050719 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-log\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050776 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d9ea68-a957-4056-867d-18e0535bec60-ovn-controller-tls-certs\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050832 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2d9ea68-a957-4056-867d-18e0535bec60-scripts\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050868 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-run\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050935 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24pwb\" (UniqueName: \"kubernetes.io/projected/d2d9ea68-a957-4056-867d-18e0535bec60-kube-api-access-24pwb\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.050964 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-lib\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.051018 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-etc-ovs\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.051053 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d9ea68-a957-4056-867d-18e0535bec60-combined-ca-bundle\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.051082 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-run-ovn\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.051109 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlp54\" (UniqueName: \"kubernetes.io/projected/a8725f3f-36cb-4074-abc0-44031fb37d60-kube-api-access-tlp54\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.051143 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8725f3f-36cb-4074-abc0-44031fb37d60-scripts\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.051204 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-log-ovn\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153081 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-run\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153129 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-log\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153154 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d9ea68-a957-4056-867d-18e0535bec60-ovn-controller-tls-certs\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153182 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2d9ea68-a957-4056-867d-18e0535bec60-scripts\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153204 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-run\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153237 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24pwb\" (UniqueName: \"kubernetes.io/projected/d2d9ea68-a957-4056-867d-18e0535bec60-kube-api-access-24pwb\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153256 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-lib\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153284 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-etc-ovs\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153306 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d9ea68-a957-4056-867d-18e0535bec60-combined-ca-bundle\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153327 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-run-ovn\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153346 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlp54\" (UniqueName: \"kubernetes.io/projected/a8725f3f-36cb-4074-abc0-44031fb37d60-kube-api-access-tlp54\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153369 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8725f3f-36cb-4074-abc0-44031fb37d60-scripts\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153402 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-log-ovn\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153722 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-run\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153797 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-log-ovn\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153802 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-run\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153890 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-log\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.153929 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2d9ea68-a957-4056-867d-18e0535bec60-var-run-ovn\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.154093 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-var-lib\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.154094 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a8725f3f-36cb-4074-abc0-44031fb37d60-etc-ovs\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.156144 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2d9ea68-a957-4056-867d-18e0535bec60-scripts\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.156464 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8725f3f-36cb-4074-abc0-44031fb37d60-scripts\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.159344 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d9ea68-a957-4056-867d-18e0535bec60-ovn-controller-tls-certs\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.160938 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d9ea68-a957-4056-867d-18e0535bec60-combined-ca-bundle\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.179041 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24pwb\" (UniqueName: \"kubernetes.io/projected/d2d9ea68-a957-4056-867d-18e0535bec60-kube-api-access-24pwb\") pod \"ovn-controller-fp7wl\" (UID: \"d2d9ea68-a957-4056-867d-18e0535bec60\") " pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.185301 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlp54\" (UniqueName: \"kubernetes.io/projected/a8725f3f-36cb-4074-abc0-44031fb37d60-kube-api-access-tlp54\") pod \"ovn-controller-ovs-jd2lt\" (UID: \"a8725f3f-36cb-4074-abc0-44031fb37d60\") " pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.273336 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:35 crc kubenswrapper[4710]: I1002 13:05:35.292930 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.975670 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.977922 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.982821 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.982864 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.983478 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-m5xzx" Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.983621 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 13:05:36 crc kubenswrapper[4710]: I1002 13:05:36.986982 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083119 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e26518b-c86d-4760-a600-b7e8a6594fab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083225 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083260 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e26518b-c86d-4760-a600-b7e8a6594fab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083335 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083384 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e26518b-c86d-4760-a600-b7e8a6594fab-config\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083442 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhdqx\" (UniqueName: \"kubernetes.io/projected/7e26518b-c86d-4760-a600-b7e8a6594fab-kube-api-access-lhdqx\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083534 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.083759 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185381 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e26518b-c86d-4760-a600-b7e8a6594fab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185442 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185484 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e26518b-c86d-4760-a600-b7e8a6594fab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185513 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185539 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e26518b-c86d-4760-a600-b7e8a6594fab-config\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185573 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhdqx\" (UniqueName: \"kubernetes.io/projected/7e26518b-c86d-4760-a600-b7e8a6594fab-kube-api-access-lhdqx\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185618 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.185668 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.186065 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e26518b-c86d-4760-a600-b7e8a6594fab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.187489 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e26518b-c86d-4760-a600-b7e8a6594fab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.187803 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.190220 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e26518b-c86d-4760-a600-b7e8a6594fab-config\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.192460 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.193237 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.193835 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e26518b-c86d-4760-a600-b7e8a6594fab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.203041 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhdqx\" (UniqueName: \"kubernetes.io/projected/7e26518b-c86d-4760-a600-b7e8a6594fab-kube-api-access-lhdqx\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.216604 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e26518b-c86d-4760-a600-b7e8a6594fab\") " pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:37 crc kubenswrapper[4710]: I1002 13:05:37.294936 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:40 crc kubenswrapper[4710]: I1002 13:05:40.972366 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bb7dbd587-vtwc9"] Oct 02 13:05:41 crc kubenswrapper[4710]: I1002 13:05:41.047385 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.369763 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.369852 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.370129 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.147:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7gqhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-784b55c5d9-mzqnv_openstack(c7858de3-1d1e-4d8e-9912-d03c99ea5ec6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:05:41 crc kubenswrapper[4710]: W1002 13:05:41.370242 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d4485f3_00c0_421a_bb40_1deb2bcad3e1.slice/crio-ffb1d83ec5fa9cbe7b2aad4370583e30231cef6aad77b882e351101e6868b80f WatchSource:0}: Error finding container ffb1d83ec5fa9cbe7b2aad4370583e30231cef6aad77b882e351101e6868b80f: Status 404 returned error can't find the container with id ffb1d83ec5fa9cbe7b2aad4370583e30231cef6aad77b882e351101e6868b80f Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.371263 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" podUID="c7858de3-1d1e-4d8e-9912-d03c99ea5ec6" Oct 02 13:05:41 crc kubenswrapper[4710]: W1002 13:05:41.375877 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod770af87f_9638_4cdb_a10d_2b40d50302ca.slice/crio-d3ce6a9121dbb564bf1407390305275d872323c79d0b0ed1dd25e300b9e61ca1 WatchSource:0}: Error finding container d3ce6a9121dbb564bf1407390305275d872323c79d0b0ed1dd25e300b9e61ca1: Status 404 returned error can't find the container with id d3ce6a9121dbb564bf1407390305275d872323c79d0b0ed1dd25e300b9e61ca1 Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.450593 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.450648 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.450840 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.147:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fl859,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-bf56b5889-6j2b4_openstack(a7d7b97d-c9fb-4d08-8c34-df63eefd96a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:05:41 crc kubenswrapper[4710]: E1002 13:05:41.456178 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" podUID="a7d7b97d-c9fb-4d08-8c34-df63eefd96a7" Oct 02 13:05:41 crc kubenswrapper[4710]: I1002 13:05:41.880977 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 13:05:41 crc kubenswrapper[4710]: I1002 13:05:41.900339 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" event={"ID":"4d4485f3-00c0-421a-bb40-1deb2bcad3e1","Type":"ContainerStarted","Data":"ffb1d83ec5fa9cbe7b2aad4370583e30231cef6aad77b882e351101e6868b80f"} Oct 02 13:05:41 crc kubenswrapper[4710]: I1002 13:05:41.901695 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"770af87f-9638-4cdb-a10d-2b40d50302ca","Type":"ContainerStarted","Data":"d3ce6a9121dbb564bf1407390305275d872323c79d0b0ed1dd25e300b9e61ca1"} Oct 02 13:05:41 crc kubenswrapper[4710]: W1002 13:05:41.908937 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podada9e73f_75af_4b4a_8e52_af2d5b8c3728.slice/crio-3dd6a3ecddc09077443d5b990d44806fe3c47305b78c019318ef8940221ebdc6 WatchSource:0}: Error finding container 3dd6a3ecddc09077443d5b990d44806fe3c47305b78c019318ef8940221ebdc6: Status 404 returned error can't find the container with id 3dd6a3ecddc09077443d5b990d44806fe3c47305b78c019318ef8940221ebdc6 Oct 02 13:05:41 crc kubenswrapper[4710]: I1002 13:05:41.982687 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 13:05:42 crc kubenswrapper[4710]: W1002 13:05:42.000448 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5d102b1_889f_487b_be09_725834944ff6.slice/crio-c85efa77e4843435cc7d0ab379ccd1705e52c1ce0ba8fd1460ff2d1a7d4b15a8 WatchSource:0}: Error finding container c85efa77e4843435cc7d0ab379ccd1705e52c1ce0ba8fd1460ff2d1a7d4b15a8: Status 404 returned error can't find the container with id c85efa77e4843435cc7d0ab379ccd1705e52c1ce0ba8fd1460ff2d1a7d4b15a8 Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.000637 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 13:05:42 crc kubenswrapper[4710]: W1002 13:05:42.020344 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40dcfd31_21c8_4581_b2f6_889737eea66d.slice/crio-4c9716f61508a96fd0dc4f12df8ded86b8890d5d23b742fd66ad2edd737066c6 WatchSource:0}: Error finding container 4c9716f61508a96fd0dc4f12df8ded86b8890d5d23b742fd66ad2edd737066c6: Status 404 returned error can't find the container with id 4c9716f61508a96fd0dc4f12df8ded86b8890d5d23b742fd66ad2edd737066c6 Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.130965 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.467868 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.493557 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.499358 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fp7wl"] Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.760257 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.924398 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"98c2f825-d5c6-47e7-a96d-8cf7227b0a24","Type":"ContainerStarted","Data":"c5701516bffa633b2b40cf750ee2c5600a244e1210b25a417d899ad28922cc57"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.924442 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"40dcfd31-21c8-4581-b2f6-889737eea66d","Type":"ContainerStarted","Data":"4c9716f61508a96fd0dc4f12df8ded86b8890d5d23b742fd66ad2edd737066c6"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.924822 4710 generic.go:334] "Generic (PLEG): container finished" podID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerID="d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11" exitCode=0 Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.924855 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" event={"ID":"4d4485f3-00c0-421a-bb40-1deb2bcad3e1","Type":"ContainerDied","Data":"d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.934370 4710 generic.go:334] "Generic (PLEG): container finished" podID="0dfff010-9bbd-4c5d-846d-50e39fe0e12e" containerID="1c4f0902d69f15c6958010fef53dfd09bb50379077c357dce49fbb0f4af59e3e" exitCode=0 Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.934463 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7985db964c-92ssp" event={"ID":"0dfff010-9bbd-4c5d-846d-50e39fe0e12e","Type":"ContainerDied","Data":"1c4f0902d69f15c6958010fef53dfd09bb50379077c357dce49fbb0f4af59e3e"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.936657 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ada9e73f-75af-4b4a-8e52-af2d5b8c3728","Type":"ContainerStarted","Data":"3dd6a3ecddc09077443d5b990d44806fe3c47305b78c019318ef8940221ebdc6"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.939110 4710 generic.go:334] "Generic (PLEG): container finished" podID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerID="8ea18dc44135f2ea4db12b7d5e2b702a0d7ec4ca3ac7aa07695253d59eb9c5f2" exitCode=0 Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.939162 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" event={"ID":"5ed3b611-ed83-4cce-843e-b64b21afecb6","Type":"ContainerDied","Data":"8ea18dc44135f2ea4db12b7d5e2b702a0d7ec4ca3ac7aa07695253d59eb9c5f2"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.941057 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c5d102b1-889f-487b-be09-725834944ff6","Type":"ContainerStarted","Data":"c85efa77e4843435cc7d0ab379ccd1705e52c1ce0ba8fd1460ff2d1a7d4b15a8"} Oct 02 13:05:42 crc kubenswrapper[4710]: I1002 13:05:42.988896 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 13:05:43 crc kubenswrapper[4710]: W1002 13:05:43.401073 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27670d96_0e34_41ae_a54c_2bb7f19b1c6e.slice/crio-6a37868078759f9356052880e4f52e77085a7da9dd5d161981019de284561ea1 WatchSource:0}: Error finding container 6a37868078759f9356052880e4f52e77085a7da9dd5d161981019de284561ea1: Status 404 returned error can't find the container with id 6a37868078759f9356052880e4f52e77085a7da9dd5d161981019de284561ea1 Oct 02 13:05:43 crc kubenswrapper[4710]: W1002 13:05:43.411025 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e26518b_c86d_4760_a600_b7e8a6594fab.slice/crio-4e8ef1a1127e691c30b998fc002d410b7e846b00d0f028ba4f0e939b6625d731 WatchSource:0}: Error finding container 4e8ef1a1127e691c30b998fc002d410b7e846b00d0f028ba4f0e939b6625d731: Status 404 returned error can't find the container with id 4e8ef1a1127e691c30b998fc002d410b7e846b00d0f028ba4f0e939b6625d731 Oct 02 13:05:43 crc kubenswrapper[4710]: W1002 13:05:43.412542 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2d9ea68_a957_4056_867d_18e0535bec60.slice/crio-cdba202804b5d8e7275bed03c8b56b7b59f5ec45979d58a523060b544155d202 WatchSource:0}: Error finding container cdba202804b5d8e7275bed03c8b56b7b59f5ec45979d58a523060b544155d202: Status 404 returned error can't find the container with id cdba202804b5d8e7275bed03c8b56b7b59f5ec45979d58a523060b544155d202 Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.478828 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.496079 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.507729 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.637262 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653494 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-dns-svc\") pod \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653578 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl859\" (UniqueName: \"kubernetes.io/projected/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-kube-api-access-fl859\") pod \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653610 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-config\") pod \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\" (UID: \"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653632 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-config\") pod \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653665 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gqhc\" (UniqueName: \"kubernetes.io/projected/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-kube-api-access-7gqhc\") pod \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\" (UID: \"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653691 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-config\") pod \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653737 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8pth\" (UniqueName: \"kubernetes.io/projected/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-kube-api-access-s8pth\") pod \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.653840 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-dns-svc\") pod \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\" (UID: \"0dfff010-9bbd-4c5d-846d-50e39fe0e12e\") " Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.654655 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-config" (OuterVolumeSpecName: "config") pod "c7858de3-1d1e-4d8e-9912-d03c99ea5ec6" (UID: "c7858de3-1d1e-4d8e-9912-d03c99ea5ec6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.655186 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7d7b97d-c9fb-4d08-8c34-df63eefd96a7" (UID: "a7d7b97d-c9fb-4d08-8c34-df63eefd96a7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.655662 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-config" (OuterVolumeSpecName: "config") pod "a7d7b97d-c9fb-4d08-8c34-df63eefd96a7" (UID: "a7d7b97d-c9fb-4d08-8c34-df63eefd96a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.659884 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-kube-api-access-7gqhc" (OuterVolumeSpecName: "kube-api-access-7gqhc") pod "c7858de3-1d1e-4d8e-9912-d03c99ea5ec6" (UID: "c7858de3-1d1e-4d8e-9912-d03c99ea5ec6"). InnerVolumeSpecName "kube-api-access-7gqhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.666001 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-kube-api-access-fl859" (OuterVolumeSpecName: "kube-api-access-fl859") pod "a7d7b97d-c9fb-4d08-8c34-df63eefd96a7" (UID: "a7d7b97d-c9fb-4d08-8c34-df63eefd96a7"). InnerVolumeSpecName "kube-api-access-fl859". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.676478 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-kube-api-access-s8pth" (OuterVolumeSpecName: "kube-api-access-s8pth") pod "0dfff010-9bbd-4c5d-846d-50e39fe0e12e" (UID: "0dfff010-9bbd-4c5d-846d-50e39fe0e12e"). InnerVolumeSpecName "kube-api-access-s8pth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.677269 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-config" (OuterVolumeSpecName: "config") pod "0dfff010-9bbd-4c5d-846d-50e39fe0e12e" (UID: "0dfff010-9bbd-4c5d-846d-50e39fe0e12e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.683868 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0dfff010-9bbd-4c5d-846d-50e39fe0e12e" (UID: "0dfff010-9bbd-4c5d-846d-50e39fe0e12e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.755969 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8pth\" (UniqueName: \"kubernetes.io/projected/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-kube-api-access-s8pth\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756008 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756020 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756032 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl859\" (UniqueName: \"kubernetes.io/projected/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-kube-api-access-fl859\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756044 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756056 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756067 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gqhc\" (UniqueName: \"kubernetes.io/projected/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6-kube-api-access-7gqhc\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.756079 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dfff010-9bbd-4c5d-846d-50e39fe0e12e-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.825179 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jd2lt"] Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.955027 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl" event={"ID":"d2d9ea68-a957-4056-867d-18e0535bec60","Type":"ContainerStarted","Data":"cdba202804b5d8e7275bed03c8b56b7b59f5ec45979d58a523060b544155d202"} Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.973949 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7985db964c-92ssp" event={"ID":"0dfff010-9bbd-4c5d-846d-50e39fe0e12e","Type":"ContainerDied","Data":"853998f71a7f3e7b61f1974e58415e2eb907f0c7b0c50df8d47aaf9714322936"} Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.974011 4710 scope.go:117] "RemoveContainer" containerID="1c4f0902d69f15c6958010fef53dfd09bb50379077c357dce49fbb0f4af59e3e" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.974037 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7985db964c-92ssp" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.976443 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"27670d96-0e34-41ae-a54c-2bb7f19b1c6e","Type":"ContainerStarted","Data":"6a37868078759f9356052880e4f52e77085a7da9dd5d161981019de284561ea1"} Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.979951 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" event={"ID":"a7d7b97d-c9fb-4d08-8c34-df63eefd96a7","Type":"ContainerDied","Data":"65e867df1725a262bcb391379fc9111afbee5429621273d5b163c8acff61a695"} Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.979968 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf56b5889-6j2b4" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.981149 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" event={"ID":"c7858de3-1d1e-4d8e-9912-d03c99ea5ec6","Type":"ContainerDied","Data":"c8c0f361ae2eaec06461213937dec82f1c3cd55b10345321d5cc625b116cbff6"} Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.981166 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784b55c5d9-mzqnv" Oct 02 13:05:43 crc kubenswrapper[4710]: I1002 13:05:43.982808 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7e26518b-c86d-4760-a600-b7e8a6594fab","Type":"ContainerStarted","Data":"4e8ef1a1127e691c30b998fc002d410b7e846b00d0f028ba4f0e939b6625d731"} Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.037839 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7985db964c-92ssp"] Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.039712 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7985db964c-92ssp"] Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.066158 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bf56b5889-6j2b4"] Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.074325 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bf56b5889-6j2b4"] Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.096631 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784b55c5d9-mzqnv"] Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.104027 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784b55c5d9-mzqnv"] Oct 02 13:05:44 crc kubenswrapper[4710]: W1002 13:05:44.855023 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7fb5713_6a35_4676_b466_6a43a4037f7b.slice/crio-1f0998615266bdef8116a38c01758db832eda01855d14331f8f6db1ae9099f7e WatchSource:0}: Error finding container 1f0998615266bdef8116a38c01758db832eda01855d14331f8f6db1ae9099f7e: Status 404 returned error can't find the container with id 1f0998615266bdef8116a38c01758db832eda01855d14331f8f6db1ae9099f7e Oct 02 13:05:44 crc kubenswrapper[4710]: W1002 13:05:44.863341 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8530440e_9b9d_40bc_95ac_b7c40899b0cb.slice/crio-547830786af86977452d67e2d13a17b0a09dda336890f14773836cf91ac87bb6 WatchSource:0}: Error finding container 547830786af86977452d67e2d13a17b0a09dda336890f14773836cf91ac87bb6: Status 404 returned error can't find the container with id 547830786af86977452d67e2d13a17b0a09dda336890f14773836cf91ac87bb6 Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.913585 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dfff010-9bbd-4c5d-846d-50e39fe0e12e" path="/var/lib/kubelet/pods/0dfff010-9bbd-4c5d-846d-50e39fe0e12e/volumes" Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.914230 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d7b97d-c9fb-4d08-8c34-df63eefd96a7" path="/var/lib/kubelet/pods/a7d7b97d-c9fb-4d08-8c34-df63eefd96a7/volumes" Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.914633 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7858de3-1d1e-4d8e-9912-d03c99ea5ec6" path="/var/lib/kubelet/pods/c7858de3-1d1e-4d8e-9912-d03c99ea5ec6/volumes" Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.990618 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerStarted","Data":"547830786af86977452d67e2d13a17b0a09dda336890f14773836cf91ac87bb6"} Oct 02 13:05:44 crc kubenswrapper[4710]: I1002 13:05:44.991635 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7fb5713-6a35-4676-b466-6a43a4037f7b","Type":"ContainerStarted","Data":"1f0998615266bdef8116a38c01758db832eda01855d14331f8f6db1ae9099f7e"} Oct 02 13:05:45 crc kubenswrapper[4710]: W1002 13:05:45.003014 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbac60db_a726_44d9_998b_cfca036d65d9.slice/crio-fde5f99ab3d0769faa84d964c3276b31923146837aa67a6211d690ea91992add WatchSource:0}: Error finding container fde5f99ab3d0769faa84d964c3276b31923146837aa67a6211d690ea91992add: Status 404 returned error can't find the container with id fde5f99ab3d0769faa84d964c3276b31923146837aa67a6211d690ea91992add Oct 02 13:05:45 crc kubenswrapper[4710]: W1002 13:05:45.007339 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8725f3f_36cb_4074_abc0_44031fb37d60.slice/crio-57dab9f19fac828fe8b9f333f902c43168f25a5cecd47638ed1cd916ad601d4b WatchSource:0}: Error finding container 57dab9f19fac828fe8b9f333f902c43168f25a5cecd47638ed1cd916ad601d4b: Status 404 returned error can't find the container with id 57dab9f19fac828fe8b9f333f902c43168f25a5cecd47638ed1cd916ad601d4b Oct 02 13:05:46 crc kubenswrapper[4710]: I1002 13:05:46.001460 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jd2lt" event={"ID":"a8725f3f-36cb-4074-abc0-44031fb37d60","Type":"ContainerStarted","Data":"57dab9f19fac828fe8b9f333f902c43168f25a5cecd47638ed1cd916ad601d4b"} Oct 02 13:05:46 crc kubenswrapper[4710]: I1002 13:05:46.002933 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cbac60db-a726-44d9-998b-cfca036d65d9","Type":"ContainerStarted","Data":"fde5f99ab3d0769faa84d964c3276b31923146837aa67a6211d690ea91992add"} Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.066820 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"40dcfd31-21c8-4581-b2f6-889737eea66d","Type":"ContainerStarted","Data":"0afdb90a0747361ddbd973f7dfb60fda044459a6f19a94167316f22ce4a07115"} Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.070959 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" event={"ID":"4d4485f3-00c0-421a-bb40-1deb2bcad3e1","Type":"ContainerStarted","Data":"7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c"} Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.071096 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.074603 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ada9e73f-75af-4b4a-8e52-af2d5b8c3728","Type":"ContainerStarted","Data":"5c069c17469a50d88b400ff6883590474afa4d3f4072f40e9cce78d262fafd5b"} Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.082608 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" event={"ID":"5ed3b611-ed83-4cce-843e-b64b21afecb6","Type":"ContainerStarted","Data":"ff9a6cfda077a6902941ec419df45348aeaaf0118df2e2798a564f13c53365b6"} Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.082776 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.088211 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c5d102b1-889f-487b-be09-725834944ff6","Type":"ContainerStarted","Data":"46481eac188bfb2fb78cbbcc379b67d9cf4d0e5169fab5abcbde70ec54ea2901"} Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.088435 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.247680 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.687327559 podStartE2EDuration="19.247665336s" podCreationTimestamp="2025-10-02 13:05:28 +0000 UTC" firstStartedPulling="2025-10-02 13:05:42.00932609 +0000 UTC m=+986.115736973" lastFinishedPulling="2025-10-02 13:05:46.569663867 +0000 UTC m=+990.676074750" observedRunningTime="2025-10-02 13:05:47.243972814 +0000 UTC m=+991.350383707" watchObservedRunningTime="2025-10-02 13:05:47.247665336 +0000 UTC m=+991.354076219" Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.271634 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" podStartSLOduration=12.648333811 podStartE2EDuration="24.271612824s" podCreationTimestamp="2025-10-02 13:05:23 +0000 UTC" firstStartedPulling="2025-10-02 13:05:29.976665694 +0000 UTC m=+974.083076577" lastFinishedPulling="2025-10-02 13:05:41.599944697 +0000 UTC m=+985.706355590" observedRunningTime="2025-10-02 13:05:47.260776483 +0000 UTC m=+991.367187376" watchObservedRunningTime="2025-10-02 13:05:47.271612824 +0000 UTC m=+991.378023707" Oct 02 13:05:47 crc kubenswrapper[4710]: I1002 13:05:47.280727 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" podStartSLOduration=23.849934524 podStartE2EDuration="24.280711031s" podCreationTimestamp="2025-10-02 13:05:23 +0000 UTC" firstStartedPulling="2025-10-02 13:05:41.372708963 +0000 UTC m=+985.479119846" lastFinishedPulling="2025-10-02 13:05:41.80348547 +0000 UTC m=+985.909896353" observedRunningTime="2025-10-02 13:05:47.279231994 +0000 UTC m=+991.385642877" watchObservedRunningTime="2025-10-02 13:05:47.280711031 +0000 UTC m=+991.387121914" Oct 02 13:05:49 crc kubenswrapper[4710]: I1002 13:05:49.105035 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"98c2f825-d5c6-47e7-a96d-8cf7227b0a24","Type":"ContainerStarted","Data":"11d9d62f9794dafce40bfec1992b6e85df6f088b226022fb0e13a8880b0312d7"} Oct 02 13:05:49 crc kubenswrapper[4710]: I1002 13:05:49.110080 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7fb5713-6a35-4676-b466-6a43a4037f7b","Type":"ContainerStarted","Data":"6987be090ed6cf0c897ec0e533b9d01229b7de9b3109dc7f672c9f5516cfef82"} Oct 02 13:05:49 crc kubenswrapper[4710]: I1002 13:05:49.111771 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"770af87f-9638-4cdb-a10d-2b40d50302ca","Type":"ContainerStarted","Data":"6b868c2abc568ca5d3358f49917fba3e3214d607ef5d3ce34983a6c316e15ab3"} Oct 02 13:05:52 crc kubenswrapper[4710]: I1002 13:05:52.176339 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cbac60db-a726-44d9-998b-cfca036d65d9","Type":"ContainerStarted","Data":"5c45a3588f20db4fab965e8a9f8acdee84120c3780f8fbc9de1c1302ec0f5b37"} Oct 02 13:05:52 crc kubenswrapper[4710]: I1002 13:05:52.530363 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:05:52 crc kubenswrapper[4710]: I1002 13:05:52.530422 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.187808 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl" event={"ID":"d2d9ea68-a957-4056-867d-18e0535bec60","Type":"ContainerStarted","Data":"478cce2596ae22d8b8ca4552d107d3fb5bf2d719b32e48b6d640631dfa56c234"} Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.187951 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-fp7wl" Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.189272 4710 generic.go:334] "Generic (PLEG): container finished" podID="a8725f3f-36cb-4074-abc0-44031fb37d60" containerID="cd35bc979e7c44d2cfaee557f5f615585a0282e200d223f2604e2ce409b0fcc7" exitCode=0 Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.189355 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jd2lt" event={"ID":"a8725f3f-36cb-4074-abc0-44031fb37d60","Type":"ContainerDied","Data":"cd35bc979e7c44d2cfaee557f5f615585a0282e200d223f2604e2ce409b0fcc7"} Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.191664 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"27670d96-0e34-41ae-a54c-2bb7f19b1c6e","Type":"ContainerStarted","Data":"35739b19eacea9fcf2439673fc3c3a190f8a8a108b09dc5fe1742feecb333446"} Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.191805 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.195800 4710 generic.go:334] "Generic (PLEG): container finished" podID="ada9e73f-75af-4b4a-8e52-af2d5b8c3728" containerID="5c069c17469a50d88b400ff6883590474afa4d3f4072f40e9cce78d262fafd5b" exitCode=0 Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.195875 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ada9e73f-75af-4b4a-8e52-af2d5b8c3728","Type":"ContainerDied","Data":"5c069c17469a50d88b400ff6883590474afa4d3f4072f40e9cce78d262fafd5b"} Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.198805 4710 generic.go:334] "Generic (PLEG): container finished" podID="40dcfd31-21c8-4581-b2f6-889737eea66d" containerID="0afdb90a0747361ddbd973f7dfb60fda044459a6f19a94167316f22ce4a07115" exitCode=0 Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.198880 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"40dcfd31-21c8-4581-b2f6-889737eea66d","Type":"ContainerDied","Data":"0afdb90a0747361ddbd973f7dfb60fda044459a6f19a94167316f22ce4a07115"} Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.203765 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7e26518b-c86d-4760-a600-b7e8a6594fab","Type":"ContainerStarted","Data":"6ae6092c5c0f0bbcb308639450ffb66b73e9a8580bd5d7ce6fbd4308aa10383d"} Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.235777 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fp7wl" podStartSLOduration=11.210757176 podStartE2EDuration="19.235754894s" podCreationTimestamp="2025-10-02 13:05:34 +0000 UTC" firstStartedPulling="2025-10-02 13:05:43.415364726 +0000 UTC m=+987.521775609" lastFinishedPulling="2025-10-02 13:05:51.440362444 +0000 UTC m=+995.546773327" observedRunningTime="2025-10-02 13:05:53.20754854 +0000 UTC m=+997.313959423" watchObservedRunningTime="2025-10-02 13:05:53.235754894 +0000 UTC m=+997.342165777" Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.289670 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.728258538 podStartE2EDuration="24.289606338s" podCreationTimestamp="2025-10-02 13:05:29 +0000 UTC" firstStartedPulling="2025-10-02 13:05:43.404458084 +0000 UTC m=+987.510868967" lastFinishedPulling="2025-10-02 13:05:51.965805884 +0000 UTC m=+996.072216767" observedRunningTime="2025-10-02 13:05:53.282479571 +0000 UTC m=+997.388890454" watchObservedRunningTime="2025-10-02 13:05:53.289606338 +0000 UTC m=+997.396018042" Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.401695 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 13:05:53 crc kubenswrapper[4710]: I1002 13:05:53.612889 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.009321 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.065102 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6b645ddc-l5f42"] Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.293267 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jd2lt" event={"ID":"a8725f3f-36cb-4074-abc0-44031fb37d60","Type":"ContainerStarted","Data":"1f73fe46448a1548526f705d70bd67fe929152f2d4dbbd094aff9e6fd2f2af7a"} Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.340523 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ada9e73f-75af-4b4a-8e52-af2d5b8c3728","Type":"ContainerStarted","Data":"40aed1f72d4302af074ac390a90839186573a04cae947c98294aadf53ca06508"} Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.345075 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerName="dnsmasq-dns" containerID="cri-o://ff9a6cfda077a6902941ec419df45348aeaaf0118df2e2798a564f13c53365b6" gracePeriod=10 Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.345310 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"40dcfd31-21c8-4581-b2f6-889737eea66d","Type":"ContainerStarted","Data":"801af715945dd81a971449eb8d9f81370733ddcd51dc0b073643dd29c118274c"} Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.387901 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=23.711216299 podStartE2EDuration="28.387884081s" podCreationTimestamp="2025-10-02 13:05:26 +0000 UTC" firstStartedPulling="2025-10-02 13:05:41.91603157 +0000 UTC m=+986.022442453" lastFinishedPulling="2025-10-02 13:05:46.592699352 +0000 UTC m=+990.699110235" observedRunningTime="2025-10-02 13:05:54.383481641 +0000 UTC m=+998.489892534" watchObservedRunningTime="2025-10-02 13:05:54.387884081 +0000 UTC m=+998.494294964" Oct 02 13:05:54 crc kubenswrapper[4710]: I1002 13:05:54.405420 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.832566639 podStartE2EDuration="28.405398608s" podCreationTimestamp="2025-10-02 13:05:26 +0000 UTC" firstStartedPulling="2025-10-02 13:05:42.02494813 +0000 UTC m=+986.131359013" lastFinishedPulling="2025-10-02 13:05:46.597780079 +0000 UTC m=+990.704190982" observedRunningTime="2025-10-02 13:05:54.399957782 +0000 UTC m=+998.506368665" watchObservedRunningTime="2025-10-02 13:05:54.405398608 +0000 UTC m=+998.511809491" Oct 02 13:05:55 crc kubenswrapper[4710]: I1002 13:05:55.357804 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerStarted","Data":"3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e"} Oct 02 13:05:55 crc kubenswrapper[4710]: I1002 13:05:55.364786 4710 generic.go:334] "Generic (PLEG): container finished" podID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerID="ff9a6cfda077a6902941ec419df45348aeaaf0118df2e2798a564f13c53365b6" exitCode=0 Oct 02 13:05:55 crc kubenswrapper[4710]: I1002 13:05:55.364837 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" event={"ID":"5ed3b611-ed83-4cce-843e-b64b21afecb6","Type":"ContainerDied","Data":"ff9a6cfda077a6902941ec419df45348aeaaf0118df2e2798a564f13c53365b6"} Oct 02 13:05:55 crc kubenswrapper[4710]: I1002 13:05:55.979626 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.141035 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-config\") pod \"5ed3b611-ed83-4cce-843e-b64b21afecb6\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.141078 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-dns-svc\") pod \"5ed3b611-ed83-4cce-843e-b64b21afecb6\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.141139 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5lnd\" (UniqueName: \"kubernetes.io/projected/5ed3b611-ed83-4cce-843e-b64b21afecb6-kube-api-access-w5lnd\") pod \"5ed3b611-ed83-4cce-843e-b64b21afecb6\" (UID: \"5ed3b611-ed83-4cce-843e-b64b21afecb6\") " Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.146675 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ed3b611-ed83-4cce-843e-b64b21afecb6-kube-api-access-w5lnd" (OuterVolumeSpecName: "kube-api-access-w5lnd") pod "5ed3b611-ed83-4cce-843e-b64b21afecb6" (UID: "5ed3b611-ed83-4cce-843e-b64b21afecb6"). InnerVolumeSpecName "kube-api-access-w5lnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.182334 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ed3b611-ed83-4cce-843e-b64b21afecb6" (UID: "5ed3b611-ed83-4cce-843e-b64b21afecb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.182444 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-config" (OuterVolumeSpecName: "config") pod "5ed3b611-ed83-4cce-843e-b64b21afecb6" (UID: "5ed3b611-ed83-4cce-843e-b64b21afecb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.242872 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.242909 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ed3b611-ed83-4cce-843e-b64b21afecb6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.242933 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5lnd\" (UniqueName: \"kubernetes.io/projected/5ed3b611-ed83-4cce-843e-b64b21afecb6-kube-api-access-w5lnd\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.376200 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jd2lt" event={"ID":"a8725f3f-36cb-4074-abc0-44031fb37d60","Type":"ContainerStarted","Data":"3e252627420e3b22e0fe26ecd87b238456d2e09418b489ca7024a900648d6b87"} Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.377182 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.377244 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.380155 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" event={"ID":"5ed3b611-ed83-4cce-843e-b64b21afecb6","Type":"ContainerDied","Data":"b60ac86f16f030076ab4f2a50eb0269d443796bbaf9f2507c219cd90c40829a4"} Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.380226 4710 scope.go:117] "RemoveContainer" containerID="ff9a6cfda077a6902941ec419df45348aeaaf0118df2e2798a564f13c53365b6" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.380473 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6b645ddc-l5f42" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.390887 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7e26518b-c86d-4760-a600-b7e8a6594fab","Type":"ContainerStarted","Data":"074f02951f66267e16f975dd2a7cf13e675b1325b21983272ac4d29647765c1e"} Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.394046 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cbac60db-a726-44d9-998b-cfca036d65d9","Type":"ContainerStarted","Data":"c8a68adf6526c700575883b4da29717d02c5da7a8f07709db686b961101196ec"} Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.412335 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jd2lt" podStartSLOduration=16.236888915 podStartE2EDuration="22.4123188s" podCreationTimestamp="2025-10-02 13:05:34 +0000 UTC" firstStartedPulling="2025-10-02 13:05:45.009797269 +0000 UTC m=+989.116208162" lastFinishedPulling="2025-10-02 13:05:51.185227164 +0000 UTC m=+995.291638047" observedRunningTime="2025-10-02 13:05:56.407888989 +0000 UTC m=+1000.514299872" watchObservedRunningTime="2025-10-02 13:05:56.4123188 +0000 UTC m=+1000.518729683" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.425338 4710 scope.go:117] "RemoveContainer" containerID="8ea18dc44135f2ea4db12b7d5e2b702a0d7ec4ca3ac7aa07695253d59eb9c5f2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.435764 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.856785879 podStartE2EDuration="21.435725044s" podCreationTimestamp="2025-10-02 13:05:35 +0000 UTC" firstStartedPulling="2025-10-02 13:05:43.41349161 +0000 UTC m=+987.519902493" lastFinishedPulling="2025-10-02 13:05:55.992430775 +0000 UTC m=+1000.098841658" observedRunningTime="2025-10-02 13:05:56.433495688 +0000 UTC m=+1000.539906571" watchObservedRunningTime="2025-10-02 13:05:56.435725044 +0000 UTC m=+1000.542135927" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.454660 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6b645ddc-l5f42"] Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.460390 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6b645ddc-l5f42"] Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.480802 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.511045274 podStartE2EDuration="24.480774449s" podCreationTimestamp="2025-10-02 13:05:32 +0000 UTC" firstStartedPulling="2025-10-02 13:05:45.00585815 +0000 UTC m=+989.112269033" lastFinishedPulling="2025-10-02 13:05:55.975587325 +0000 UTC m=+1000.081998208" observedRunningTime="2025-10-02 13:05:56.475567939 +0000 UTC m=+1000.581978842" watchObservedRunningTime="2025-10-02 13:05:56.480774449 +0000 UTC m=+1000.587185332" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.863194 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-j7xs2"] Oct 02 13:05:56 crc kubenswrapper[4710]: E1002 13:05:56.863507 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerName="init" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.863521 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerName="init" Oct 02 13:05:56 crc kubenswrapper[4710]: E1002 13:05:56.863532 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerName="dnsmasq-dns" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.863538 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerName="dnsmasq-dns" Oct 02 13:05:56 crc kubenswrapper[4710]: E1002 13:05:56.863554 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dfff010-9bbd-4c5d-846d-50e39fe0e12e" containerName="init" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.863561 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dfff010-9bbd-4c5d-846d-50e39fe0e12e" containerName="init" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.863700 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dfff010-9bbd-4c5d-846d-50e39fe0e12e" containerName="init" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.863720 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" containerName="dnsmasq-dns" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.864277 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.871665 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.878959 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j7xs2"] Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.916137 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ed3b611-ed83-4cce-843e-b64b21afecb6" path="/var/lib/kubelet/pods/5ed3b611-ed83-4cce-843e-b64b21afecb6/volumes" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.959803 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7250728d-33a5-4939-8fc3-19fe161d8035-config\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.959900 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7250728d-33a5-4939-8fc3-19fe161d8035-ovn-rundir\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.959933 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7250728d-33a5-4939-8fc3-19fe161d8035-ovs-rundir\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.959952 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7250728d-33a5-4939-8fc3-19fe161d8035-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.959993 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggt48\" (UniqueName: \"kubernetes.io/projected/7250728d-33a5-4939-8fc3-19fe161d8035-kube-api-access-ggt48\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.960037 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7250728d-33a5-4939-8fc3-19fe161d8035-combined-ca-bundle\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.983932 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f599bcb7-6kqdx"] Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.985460 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:56 crc kubenswrapper[4710]: I1002 13:05:56.987669 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.001118 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f599bcb7-6kqdx"] Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061173 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7250728d-33a5-4939-8fc3-19fe161d8035-ovn-rundir\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061245 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7250728d-33a5-4939-8fc3-19fe161d8035-ovs-rundir\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061300 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-config\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061326 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7250728d-33a5-4939-8fc3-19fe161d8035-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061439 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-dns-svc\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061527 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggt48\" (UniqueName: \"kubernetes.io/projected/7250728d-33a5-4939-8fc3-19fe161d8035-kube-api-access-ggt48\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061595 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7250728d-33a5-4939-8fc3-19fe161d8035-ovn-rundir\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061607 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdxgk\" (UniqueName: \"kubernetes.io/projected/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-kube-api-access-tdxgk\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061721 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061798 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7250728d-33a5-4939-8fc3-19fe161d8035-combined-ca-bundle\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061876 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7250728d-33a5-4939-8fc3-19fe161d8035-ovs-rundir\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.061953 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7250728d-33a5-4939-8fc3-19fe161d8035-config\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.062593 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7250728d-33a5-4939-8fc3-19fe161d8035-config\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.070481 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7250728d-33a5-4939-8fc3-19fe161d8035-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.070487 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7250728d-33a5-4939-8fc3-19fe161d8035-combined-ca-bundle\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.079903 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggt48\" (UniqueName: \"kubernetes.io/projected/7250728d-33a5-4939-8fc3-19fe161d8035-kube-api-access-ggt48\") pod \"ovn-controller-metrics-j7xs2\" (UID: \"7250728d-33a5-4939-8fc3-19fe161d8035\") " pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.164204 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-config\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.164275 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-dns-svc\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.164371 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdxgk\" (UniqueName: \"kubernetes.io/projected/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-kube-api-access-tdxgk\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.164410 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.165633 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-dns-svc\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.165696 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-config\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.165715 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.179291 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j7xs2" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.197589 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdxgk\" (UniqueName: \"kubernetes.io/projected/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-kube-api-access-tdxgk\") pod \"dnsmasq-dns-7f599bcb7-6kqdx\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.267281 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f599bcb7-6kqdx"] Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.269008 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.296082 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.296522 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f65f998dc-rnhk9"] Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.298373 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.300504 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.313693 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f65f998dc-rnhk9"] Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.386359 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-sb\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.386428 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4xqr\" (UniqueName: \"kubernetes.io/projected/3ca48c29-e316-4238-8b1c-d6887f34f1af-kube-api-access-b4xqr\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.386494 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-nb\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.386525 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-config\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.386575 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-dns-svc\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.488786 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-config\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.489194 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-dns-svc\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.489296 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-sb\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.489349 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4xqr\" (UniqueName: \"kubernetes.io/projected/3ca48c29-e316-4238-8b1c-d6887f34f1af-kube-api-access-b4xqr\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.489427 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-nb\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.490287 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-config\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.490849 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-sb\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.491321 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-nb\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.492726 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-dns-svc\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.513355 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4xqr\" (UniqueName: \"kubernetes.io/projected/3ca48c29-e316-4238-8b1c-d6887f34f1af-kube-api-access-b4xqr\") pod \"dnsmasq-dns-5f65f998dc-rnhk9\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.614590 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.614643 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.621527 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.799207 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f599bcb7-6kqdx"] Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.809549 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j7xs2"] Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.813499 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:57 crc kubenswrapper[4710]: I1002 13:05:57.863860 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.012127 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.012443 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 13:05:58 crc kubenswrapper[4710]: W1002 13:05:58.128663 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca48c29_e316_4238_8b1c_d6887f34f1af.slice/crio-f5491113324cf2cdede0871f9bdba26a1bc475d79146f5ae75f5192b1d06704a WatchSource:0}: Error finding container f5491113324cf2cdede0871f9bdba26a1bc475d79146f5ae75f5192b1d06704a: Status 404 returned error can't find the container with id f5491113324cf2cdede0871f9bdba26a1bc475d79146f5ae75f5192b1d06704a Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.129637 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f65f998dc-rnhk9"] Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.296087 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.337450 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.422140 4710 generic.go:334] "Generic (PLEG): container finished" podID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerID="3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1" exitCode=0 Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.422232 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" event={"ID":"3ca48c29-e316-4238-8b1c-d6887f34f1af","Type":"ContainerDied","Data":"3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1"} Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.422288 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" event={"ID":"3ca48c29-e316-4238-8b1c-d6887f34f1af","Type":"ContainerStarted","Data":"f5491113324cf2cdede0871f9bdba26a1bc475d79146f5ae75f5192b1d06704a"} Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.424210 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j7xs2" event={"ID":"7250728d-33a5-4939-8fc3-19fe161d8035","Type":"ContainerStarted","Data":"90e40c08500bb3b69bca640c9952f70d5e4fa7bc931244f3eac2eb4265c3eafa"} Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.424449 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j7xs2" event={"ID":"7250728d-33a5-4939-8fc3-19fe161d8035","Type":"ContainerStarted","Data":"fd0e36acc535123c2476068f6dcfc83a911628b2c1b16eb0cea686988c86fa5b"} Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.425484 4710 generic.go:334] "Generic (PLEG): container finished" podID="a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" containerID="26fcf70df2b93bf44a2192ceee1131e54996c6b6c0c50a0a6dbb43ea42acf41d" exitCode=0 Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.425590 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" event={"ID":"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f","Type":"ContainerDied","Data":"26fcf70df2b93bf44a2192ceee1131e54996c6b6c0c50a0a6dbb43ea42acf41d"} Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.425617 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" event={"ID":"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f","Type":"ContainerStarted","Data":"2c400ce7c747b529287cac3013bf01eb5001006019afbee31c349f04d1e20396"} Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.426777 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.488860 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.491763 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.507891 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-j7xs2" podStartSLOduration=2.5078681229999997 podStartE2EDuration="2.507868123s" podCreationTimestamp="2025-10-02 13:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:05:58.489103045 +0000 UTC m=+1002.595513928" watchObservedRunningTime="2025-10-02 13:05:58.507868123 +0000 UTC m=+1002.614279006" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.768333 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.770350 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.772913 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.773321 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.773505 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.780134 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-w9h24" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.795488 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.834159 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.914006 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.914053 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c29623cf-e442-438d-8f92-d16927045438-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.914087 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbxmg\" (UniqueName: \"kubernetes.io/projected/c29623cf-e442-438d-8f92-d16927045438-kube-api-access-jbxmg\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.914679 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c29623cf-e442-438d-8f92-d16927045438-scripts\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.914971 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.915026 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c29623cf-e442-438d-8f92-d16927045438-config\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:58 crc kubenswrapper[4710]: I1002 13:05:58.915121 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.017479 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-dns-svc\") pod \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.017537 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-ovsdbserver-nb\") pod \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.017574 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-config\") pod \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.017660 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdxgk\" (UniqueName: \"kubernetes.io/projected/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-kube-api-access-tdxgk\") pod \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\" (UID: \"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f\") " Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018105 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018166 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018192 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c29623cf-e442-438d-8f92-d16927045438-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018223 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbxmg\" (UniqueName: \"kubernetes.io/projected/c29623cf-e442-438d-8f92-d16927045438-kube-api-access-jbxmg\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018255 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c29623cf-e442-438d-8f92-d16927045438-scripts\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018304 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.018346 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c29623cf-e442-438d-8f92-d16927045438-config\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.019340 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c29623cf-e442-438d-8f92-d16927045438-config\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.020358 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c29623cf-e442-438d-8f92-d16927045438-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.020441 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c29623cf-e442-438d-8f92-d16927045438-scripts\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.025585 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.025703 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.027773 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29623cf-e442-438d-8f92-d16927045438-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.028026 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-kube-api-access-tdxgk" (OuterVolumeSpecName: "kube-api-access-tdxgk") pod "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" (UID: "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f"). InnerVolumeSpecName "kube-api-access-tdxgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.043626 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" (UID: "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.046132 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbxmg\" (UniqueName: \"kubernetes.io/projected/c29623cf-e442-438d-8f92-d16927045438-kube-api-access-jbxmg\") pod \"ovn-northd-0\" (UID: \"c29623cf-e442-438d-8f92-d16927045438\") " pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.046694 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-config" (OuterVolumeSpecName: "config") pod "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" (UID: "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.055701 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" (UID: "a62153a9-4ba0-4beb-9fc7-51cb15f29b4f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.106307 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.119560 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.119607 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.119623 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.119637 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdxgk\" (UniqueName: \"kubernetes.io/projected/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f-kube-api-access-tdxgk\") on node \"crc\" DevicePath \"\"" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.437720 4710 generic.go:334] "Generic (PLEG): container finished" podID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerID="3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e" exitCode=0 Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.437795 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerDied","Data":"3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e"} Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.443070 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" event={"ID":"3ca48c29-e316-4238-8b1c-d6887f34f1af","Type":"ContainerStarted","Data":"3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f"} Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.443923 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.448466 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" event={"ID":"a62153a9-4ba0-4beb-9fc7-51cb15f29b4f","Type":"ContainerDied","Data":"2c400ce7c747b529287cac3013bf01eb5001006019afbee31c349f04d1e20396"} Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.448553 4710 scope.go:117] "RemoveContainer" containerID="26fcf70df2b93bf44a2192ceee1131e54996c6b6c0c50a0a6dbb43ea42acf41d" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.449032 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f599bcb7-6kqdx" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.507511 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" podStartSLOduration=2.5074868329999997 podStartE2EDuration="2.507486833s" podCreationTimestamp="2025-10-02 13:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:05:59.504920589 +0000 UTC m=+1003.611331472" watchObservedRunningTime="2025-10-02 13:05:59.507486833 +0000 UTC m=+1003.613897716" Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.587973 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f599bcb7-6kqdx"] Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.609452 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f599bcb7-6kqdx"] Oct 02 13:05:59 crc kubenswrapper[4710]: I1002 13:05:59.617495 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.136932 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.255907 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f65f998dc-rnhk9"] Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.310798 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bb78f7fbf-gkh8q"] Oct 02 13:06:00 crc kubenswrapper[4710]: E1002 13:06:00.311134 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" containerName="init" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.311149 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" containerName="init" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.311312 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" containerName="init" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.312863 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.360049 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bb78f7fbf-gkh8q"] Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.457081 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-dns-svc\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.457154 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-config\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.457357 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.457482 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.457590 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfgkv\" (UniqueName: \"kubernetes.io/projected/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-kube-api-access-pfgkv\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.477618 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c29623cf-e442-438d-8f92-d16927045438","Type":"ContainerStarted","Data":"d91d4627963d38f9dd64a56faff367aeb110141b021ee834f8a67b065716841f"} Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.477655 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c29623cf-e442-438d-8f92-d16927045438","Type":"ContainerStarted","Data":"dda3935013437e9237798d9a51c52149192c3d796e47c2d61ec8edcba0846ef2"} Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.558680 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfgkv\" (UniqueName: \"kubernetes.io/projected/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-kube-api-access-pfgkv\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.558825 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-dns-svc\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.558876 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-config\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.558928 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.558986 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.559740 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.561047 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.561211 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-dns-svc\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.561282 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-config\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.579025 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfgkv\" (UniqueName: \"kubernetes.io/projected/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-kube-api-access-pfgkv\") pod \"dnsmasq-dns-7bb78f7fbf-gkh8q\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.639132 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:00 crc kubenswrapper[4710]: I1002 13:06:00.921969 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a62153a9-4ba0-4beb-9fc7-51cb15f29b4f" path="/var/lib/kubelet/pods/a62153a9-4ba0-4beb-9fc7-51cb15f29b4f/volumes" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.101495 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bb78f7fbf-gkh8q"] Oct 02 13:06:01 crc kubenswrapper[4710]: W1002 13:06:01.105729 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6af2e16f_eced_4e5f_90b8_e1fffd0d6573.slice/crio-f97ec75f287de9875b9885fa95652d7e83a483817c39d637ee0d56673a3ed9bd WatchSource:0}: Error finding container f97ec75f287de9875b9885fa95652d7e83a483817c39d637ee0d56673a3ed9bd: Status 404 returned error can't find the container with id f97ec75f287de9875b9885fa95652d7e83a483817c39d637ee0d56673a3ed9bd Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.318819 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.342406 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.369423 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.369632 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.369681 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.369941 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9gc2k" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.383431 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.482669 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/67031d3a-15f2-4b86-92d7-b525a3a8420e-cache\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.482984 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g5jc\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-kube-api-access-5g5jc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.483064 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.483086 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.483122 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/67031d3a-15f2-4b86-92d7-b525a3a8420e-lock\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.506951 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c29623cf-e442-438d-8f92-d16927045438","Type":"ContainerStarted","Data":"00384e3ff906479fb8f247a6736e63aa60b3060fd5241809c5e92216b9373e40"} Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.508444 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.514998 4710 generic.go:334] "Generic (PLEG): container finished" podID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerID="dd23a219ffc22b40a4bd6c12295fc46869251d7471604e7b434cec63651b5536" exitCode=0 Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.515112 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" event={"ID":"6af2e16f-eced-4e5f-90b8-e1fffd0d6573","Type":"ContainerDied","Data":"dd23a219ffc22b40a4bd6c12295fc46869251d7471604e7b434cec63651b5536"} Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.515175 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" event={"ID":"6af2e16f-eced-4e5f-90b8-e1fffd0d6573","Type":"ContainerStarted","Data":"f97ec75f287de9875b9885fa95652d7e83a483817c39d637ee0d56673a3ed9bd"} Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.515667 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerName="dnsmasq-dns" containerID="cri-o://3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f" gracePeriod=10 Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.534802 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.121723059 podStartE2EDuration="3.53469207s" podCreationTimestamp="2025-10-02 13:05:58 +0000 UTC" firstStartedPulling="2025-10-02 13:05:59.582547768 +0000 UTC m=+1003.688958641" lastFinishedPulling="2025-10-02 13:05:59.995516769 +0000 UTC m=+1004.101927652" observedRunningTime="2025-10-02 13:06:01.530154177 +0000 UTC m=+1005.636565060" watchObservedRunningTime="2025-10-02 13:06:01.53469207 +0000 UTC m=+1005.641102953" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.585258 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.585350 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.585409 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/67031d3a-15f2-4b86-92d7-b525a3a8420e-lock\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.585491 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/67031d3a-15f2-4b86-92d7-b525a3a8420e-cache\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: E1002 13:06:01.585510 4710 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.585550 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g5jc\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-kube-api-access-5g5jc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: E1002 13:06:01.585555 4710 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 13:06:01 crc kubenswrapper[4710]: E1002 13:06:01.585630 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift podName:67031d3a-15f2-4b86-92d7-b525a3a8420e nodeName:}" failed. No retries permitted until 2025-10-02 13:06:02.085604942 +0000 UTC m=+1006.192015825 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift") pod "swift-storage-0" (UID: "67031d3a-15f2-4b86-92d7-b525a3a8420e") : configmap "swift-ring-files" not found Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.586075 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.586850 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/67031d3a-15f2-4b86-92d7-b525a3a8420e-lock\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.587352 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/67031d3a-15f2-4b86-92d7-b525a3a8420e-cache\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.604319 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g5jc\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-kube-api-access-5g5jc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.649180 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.655278 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-xvqwj"] Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.665470 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.669533 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.669810 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.669631 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xvqwj"] Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.670400 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.770976 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.791555 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-combined-ca-bundle\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.791623 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-scripts\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.791681 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-swiftconf\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.791738 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-dispersionconf\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.791791 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-ring-data-devices\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.791887 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/989f6404-4404-4e77-b3bc-f3f27a7e7520-etc-swift\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.792077 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghtj7\" (UniqueName: \"kubernetes.io/projected/989f6404-4404-4e77-b3bc-f3f27a7e7520-kube-api-access-ghtj7\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.865366 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896538 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-combined-ca-bundle\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896616 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-scripts\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896660 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-swiftconf\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896693 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-dispersionconf\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896717 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-ring-data-devices\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896733 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/989f6404-4404-4e77-b3bc-f3f27a7e7520-etc-swift\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.896839 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghtj7\" (UniqueName: \"kubernetes.io/projected/989f6404-4404-4e77-b3bc-f3f27a7e7520-kube-api-access-ghtj7\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.903238 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-ring-data-devices\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.903533 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/989f6404-4404-4e77-b3bc-f3f27a7e7520-etc-swift\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.906730 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-scripts\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.909130 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-swiftconf\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.913019 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-combined-ca-bundle\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.913375 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-dispersionconf\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.936099 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghtj7\" (UniqueName: \"kubernetes.io/projected/989f6404-4404-4e77-b3bc-f3f27a7e7520-kube-api-access-ghtj7\") pod \"swift-ring-rebalance-xvqwj\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:01 crc kubenswrapper[4710]: I1002 13:06:01.980436 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.097270 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.102626 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-nb\") pod \"3ca48c29-e316-4238-8b1c-d6887f34f1af\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.102801 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-config\") pod \"3ca48c29-e316-4238-8b1c-d6887f34f1af\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.102846 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-dns-svc\") pod \"3ca48c29-e316-4238-8b1c-d6887f34f1af\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.102917 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4xqr\" (UniqueName: \"kubernetes.io/projected/3ca48c29-e316-4238-8b1c-d6887f34f1af-kube-api-access-b4xqr\") pod \"3ca48c29-e316-4238-8b1c-d6887f34f1af\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.102960 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-sb\") pod \"3ca48c29-e316-4238-8b1c-d6887f34f1af\" (UID: \"3ca48c29-e316-4238-8b1c-d6887f34f1af\") " Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.103392 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:02 crc kubenswrapper[4710]: E1002 13:06:02.103563 4710 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 13:06:02 crc kubenswrapper[4710]: E1002 13:06:02.103588 4710 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 13:06:02 crc kubenswrapper[4710]: E1002 13:06:02.103647 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift podName:67031d3a-15f2-4b86-92d7-b525a3a8420e nodeName:}" failed. No retries permitted until 2025-10-02 13:06:03.103630826 +0000 UTC m=+1007.210041709 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift") pod "swift-storage-0" (UID: "67031d3a-15f2-4b86-92d7-b525a3a8420e") : configmap "swift-ring-files" not found Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.108078 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca48c29-e316-4238-8b1c-d6887f34f1af-kube-api-access-b4xqr" (OuterVolumeSpecName: "kube-api-access-b4xqr") pod "3ca48c29-e316-4238-8b1c-d6887f34f1af" (UID: "3ca48c29-e316-4238-8b1c-d6887f34f1af"). InnerVolumeSpecName "kube-api-access-b4xqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.155695 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ca48c29-e316-4238-8b1c-d6887f34f1af" (UID: "3ca48c29-e316-4238-8b1c-d6887f34f1af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.156734 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ca48c29-e316-4238-8b1c-d6887f34f1af" (UID: "3ca48c29-e316-4238-8b1c-d6887f34f1af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.167942 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-config" (OuterVolumeSpecName: "config") pod "3ca48c29-e316-4238-8b1c-d6887f34f1af" (UID: "3ca48c29-e316-4238-8b1c-d6887f34f1af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.191598 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ca48c29-e316-4238-8b1c-d6887f34f1af" (UID: "3ca48c29-e316-4238-8b1c-d6887f34f1af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.205533 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.205575 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.205586 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.205600 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4xqr\" (UniqueName: \"kubernetes.io/projected/3ca48c29-e316-4238-8b1c-d6887f34f1af-kube-api-access-b4xqr\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.205614 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ca48c29-e316-4238-8b1c-d6887f34f1af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.533607 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" event={"ID":"6af2e16f-eced-4e5f-90b8-e1fffd0d6573","Type":"ContainerStarted","Data":"d069eaf80733544e7afcb6b1d6ccd56e4a4d04b44060b04ae2c7e9d2bd1c44bf"} Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.534847 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.538628 4710 generic.go:334] "Generic (PLEG): container finished" podID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerID="3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f" exitCode=0 Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.538997 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.538992 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" event={"ID":"3ca48c29-e316-4238-8b1c-d6887f34f1af","Type":"ContainerDied","Data":"3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f"} Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.539170 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f65f998dc-rnhk9" event={"ID":"3ca48c29-e316-4238-8b1c-d6887f34f1af","Type":"ContainerDied","Data":"f5491113324cf2cdede0871f9bdba26a1bc475d79146f5ae75f5192b1d06704a"} Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.539192 4710 scope.go:117] "RemoveContainer" containerID="3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.547878 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xvqwj"] Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.586454 4710 scope.go:117] "RemoveContainer" containerID="3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.624323 4710 scope.go:117] "RemoveContainer" containerID="3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f" Oct 02 13:06:02 crc kubenswrapper[4710]: E1002 13:06:02.625093 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f\": container with ID starting with 3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f not found: ID does not exist" containerID="3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.625126 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f"} err="failed to get container status \"3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f\": rpc error: code = NotFound desc = could not find container \"3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f\": container with ID starting with 3668452fa1e3de1bbf348b52424572894b9b58ea0501c5ca2f55749662d87d9f not found: ID does not exist" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.625173 4710 scope.go:117] "RemoveContainer" containerID="3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1" Oct 02 13:06:02 crc kubenswrapper[4710]: E1002 13:06:02.625777 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1\": container with ID starting with 3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1 not found: ID does not exist" containerID="3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.625807 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1"} err="failed to get container status \"3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1\": rpc error: code = NotFound desc = could not find container \"3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1\": container with ID starting with 3850c309ec7e9ff4f0becffe426d92d83e81c5c26ce0b0cbf1cad2551fc2f4d1 not found: ID does not exist" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.651628 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" podStartSLOduration=2.651608329 podStartE2EDuration="2.651608329s" podCreationTimestamp="2025-10-02 13:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:06:02.56031152 +0000 UTC m=+1006.666722433" watchObservedRunningTime="2025-10-02 13:06:02.651608329 +0000 UTC m=+1006.758019202" Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.661437 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f65f998dc-rnhk9"] Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.669806 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f65f998dc-rnhk9"] Oct 02 13:06:02 crc kubenswrapper[4710]: I1002 13:06:02.925930 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" path="/var/lib/kubelet/pods/3ca48c29-e316-4238-8b1c-d6887f34f1af/volumes" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.134542 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:03 crc kubenswrapper[4710]: E1002 13:06:03.134727 4710 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 13:06:03 crc kubenswrapper[4710]: E1002 13:06:03.134758 4710 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 13:06:03 crc kubenswrapper[4710]: E1002 13:06:03.134809 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift podName:67031d3a-15f2-4b86-92d7-b525a3a8420e nodeName:}" failed. No retries permitted until 2025-10-02 13:06:05.134795624 +0000 UTC m=+1009.241206507 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift") pod "swift-storage-0" (UID: "67031d3a-15f2-4b86-92d7-b525a3a8420e") : configmap "swift-ring-files" not found Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.547621 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xvqwj" event={"ID":"989f6404-4404-4e77-b3bc-f3f27a7e7520","Type":"ContainerStarted","Data":"87f25bcc9dccbb2ab7701ca64b9aa64f4bc41a28e8644155d8562a41088cea20"} Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.638411 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-hh65j"] Oct 02 13:06:03 crc kubenswrapper[4710]: E1002 13:06:03.638805 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerName="dnsmasq-dns" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.638822 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerName="dnsmasq-dns" Oct 02 13:06:03 crc kubenswrapper[4710]: E1002 13:06:03.638834 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerName="init" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.638841 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerName="init" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.639058 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca48c29-e316-4238-8b1c-d6887f34f1af" containerName="dnsmasq-dns" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.639637 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hh65j" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.658987 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hh65j"] Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.744526 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgx28\" (UniqueName: \"kubernetes.io/projected/afdc93b0-6465-43e9-8d1f-045f2ee24460-kube-api-access-rgx28\") pod \"glance-db-create-hh65j\" (UID: \"afdc93b0-6465-43e9-8d1f-045f2ee24460\") " pod="openstack/glance-db-create-hh65j" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.846077 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgx28\" (UniqueName: \"kubernetes.io/projected/afdc93b0-6465-43e9-8d1f-045f2ee24460-kube-api-access-rgx28\") pod \"glance-db-create-hh65j\" (UID: \"afdc93b0-6465-43e9-8d1f-045f2ee24460\") " pod="openstack/glance-db-create-hh65j" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.864117 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgx28\" (UniqueName: \"kubernetes.io/projected/afdc93b0-6465-43e9-8d1f-045f2ee24460-kube-api-access-rgx28\") pod \"glance-db-create-hh65j\" (UID: \"afdc93b0-6465-43e9-8d1f-045f2ee24460\") " pod="openstack/glance-db-create-hh65j" Oct 02 13:06:03 crc kubenswrapper[4710]: I1002 13:06:03.958776 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hh65j" Oct 02 13:06:04 crc kubenswrapper[4710]: I1002 13:06:04.410977 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hh65j"] Oct 02 13:06:04 crc kubenswrapper[4710]: I1002 13:06:04.560012 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hh65j" event={"ID":"afdc93b0-6465-43e9-8d1f-045f2ee24460","Type":"ContainerStarted","Data":"996cb67272a77c1d4aeb470c6c49ee5e2287afdd248e431f1499d3c5281b3bd4"} Oct 02 13:06:05 crc kubenswrapper[4710]: I1002 13:06:05.168622 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:05 crc kubenswrapper[4710]: E1002 13:06:05.168842 4710 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 13:06:05 crc kubenswrapper[4710]: E1002 13:06:05.168871 4710 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 13:06:05 crc kubenswrapper[4710]: E1002 13:06:05.168930 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift podName:67031d3a-15f2-4b86-92d7-b525a3a8420e nodeName:}" failed. No retries permitted until 2025-10-02 13:06:09.168910874 +0000 UTC m=+1013.275321757 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift") pod "swift-storage-0" (UID: "67031d3a-15f2-4b86-92d7-b525a3a8420e") : configmap "swift-ring-files" not found Oct 02 13:06:05 crc kubenswrapper[4710]: I1002 13:06:05.915955 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 13:06:06 crc kubenswrapper[4710]: I1002 13:06:06.004953 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 13:06:06 crc kubenswrapper[4710]: I1002 13:06:06.578468 4710 generic.go:334] "Generic (PLEG): container finished" podID="afdc93b0-6465-43e9-8d1f-045f2ee24460" containerID="efa390cf6d953d43af790b8552702466050dce517b3cee5e250e1061e5c4c296" exitCode=0 Oct 02 13:06:06 crc kubenswrapper[4710]: I1002 13:06:06.578522 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hh65j" event={"ID":"afdc93b0-6465-43e9-8d1f-045f2ee24460","Type":"ContainerDied","Data":"efa390cf6d953d43af790b8552702466050dce517b3cee5e250e1061e5c4c296"} Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.021649 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-hzh54"] Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.023573 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.029959 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hzh54"] Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.133472 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqm6z\" (UniqueName: \"kubernetes.io/projected/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51-kube-api-access-dqm6z\") pod \"keystone-db-create-hzh54\" (UID: \"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51\") " pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.235780 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqm6z\" (UniqueName: \"kubernetes.io/projected/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51-kube-api-access-dqm6z\") pod \"keystone-db-create-hzh54\" (UID: \"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51\") " pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.256120 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqm6z\" (UniqueName: \"kubernetes.io/projected/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51-kube-api-access-dqm6z\") pod \"keystone-db-create-hzh54\" (UID: \"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51\") " pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.327130 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-msp99"] Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.329351 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-msp99" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.337057 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-msp99"] Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.374895 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.439795 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66lpf\" (UniqueName: \"kubernetes.io/projected/5ed1b30c-53d5-437e-b5c2-d5526668c15d-kube-api-access-66lpf\") pod \"placement-db-create-msp99\" (UID: \"5ed1b30c-53d5-437e-b5c2-d5526668c15d\") " pod="openstack/placement-db-create-msp99" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.542329 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66lpf\" (UniqueName: \"kubernetes.io/projected/5ed1b30c-53d5-437e-b5c2-d5526668c15d-kube-api-access-66lpf\") pod \"placement-db-create-msp99\" (UID: \"5ed1b30c-53d5-437e-b5c2-d5526668c15d\") " pod="openstack/placement-db-create-msp99" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.561177 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66lpf\" (UniqueName: \"kubernetes.io/projected/5ed1b30c-53d5-437e-b5c2-d5526668c15d-kube-api-access-66lpf\") pod \"placement-db-create-msp99\" (UID: \"5ed1b30c-53d5-437e-b5c2-d5526668c15d\") " pod="openstack/placement-db-create-msp99" Oct 02 13:06:08 crc kubenswrapper[4710]: I1002 13:06:08.698091 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-msp99" Oct 02 13:06:09 crc kubenswrapper[4710]: I1002 13:06:09.259200 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:09 crc kubenswrapper[4710]: E1002 13:06:09.259386 4710 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 13:06:09 crc kubenswrapper[4710]: E1002 13:06:09.259416 4710 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 13:06:09 crc kubenswrapper[4710]: E1002 13:06:09.259490 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift podName:67031d3a-15f2-4b86-92d7-b525a3a8420e nodeName:}" failed. No retries permitted until 2025-10-02 13:06:17.259464751 +0000 UTC m=+1021.365875634 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift") pod "swift-storage-0" (UID: "67031d3a-15f2-4b86-92d7-b525a3a8420e") : configmap "swift-ring-files" not found Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.169015 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-79f7g"] Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.170409 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.176199 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-79f7g"] Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.180890 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cx2c\" (UniqueName: \"kubernetes.io/projected/53530783-8e7c-4e44-9484-6e5e9ae70bd2-kube-api-access-4cx2c\") pod \"watcher-db-create-79f7g\" (UID: \"53530783-8e7c-4e44-9484-6e5e9ae70bd2\") " pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.227023 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hh65j" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.282445 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cx2c\" (UniqueName: \"kubernetes.io/projected/53530783-8e7c-4e44-9484-6e5e9ae70bd2-kube-api-access-4cx2c\") pod \"watcher-db-create-79f7g\" (UID: \"53530783-8e7c-4e44-9484-6e5e9ae70bd2\") " pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.302674 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cx2c\" (UniqueName: \"kubernetes.io/projected/53530783-8e7c-4e44-9484-6e5e9ae70bd2-kube-api-access-4cx2c\") pod \"watcher-db-create-79f7g\" (UID: \"53530783-8e7c-4e44-9484-6e5e9ae70bd2\") " pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.367895 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.384345 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgx28\" (UniqueName: \"kubernetes.io/projected/afdc93b0-6465-43e9-8d1f-045f2ee24460-kube-api-access-rgx28\") pod \"afdc93b0-6465-43e9-8d1f-045f2ee24460\" (UID: \"afdc93b0-6465-43e9-8d1f-045f2ee24460\") " Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.391425 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdc93b0-6465-43e9-8d1f-045f2ee24460-kube-api-access-rgx28" (OuterVolumeSpecName: "kube-api-access-rgx28") pod "afdc93b0-6465-43e9-8d1f-045f2ee24460" (UID: "afdc93b0-6465-43e9-8d1f-045f2ee24460"). InnerVolumeSpecName "kube-api-access-rgx28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.486080 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgx28\" (UniqueName: \"kubernetes.io/projected/afdc93b0-6465-43e9-8d1f-045f2ee24460-kube-api-access-rgx28\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.616315 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xvqwj" event={"ID":"989f6404-4404-4e77-b3bc-f3f27a7e7520","Type":"ContainerStarted","Data":"eedb51a6b5275fb5e32aafb1717087a039bd72fd2c3e04367920aa94e4fb032e"} Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.622384 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerStarted","Data":"17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb"} Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.626012 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hh65j" event={"ID":"afdc93b0-6465-43e9-8d1f-045f2ee24460","Type":"ContainerDied","Data":"996cb67272a77c1d4aeb470c6c49ee5e2287afdd248e431f1499d3c5281b3bd4"} Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.626055 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="996cb67272a77c1d4aeb470c6c49ee5e2287afdd248e431f1499d3c5281b3bd4" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.626056 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hh65j" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.640937 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.643460 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-xvqwj" podStartSLOduration=1.95003867 podStartE2EDuration="9.643439698s" podCreationTimestamp="2025-10-02 13:06:01 +0000 UTC" firstStartedPulling="2025-10-02 13:06:02.56794811 +0000 UTC m=+1006.674358993" lastFinishedPulling="2025-10-02 13:06:10.261349138 +0000 UTC m=+1014.367760021" observedRunningTime="2025-10-02 13:06:10.633248834 +0000 UTC m=+1014.739659717" watchObservedRunningTime="2025-10-02 13:06:10.643439698 +0000 UTC m=+1014.749850581" Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.735877 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bb7dbd587-vtwc9"] Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.736409 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerName="dnsmasq-dns" containerID="cri-o://7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c" gracePeriod=10 Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.758641 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-msp99"] Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.811628 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hzh54"] Oct 02 13:06:10 crc kubenswrapper[4710]: I1002 13:06:10.883854 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-79f7g"] Oct 02 13:06:10 crc kubenswrapper[4710]: W1002 13:06:10.909395 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53530783_8e7c_4e44_9484_6e5e9ae70bd2.slice/crio-e6cfd04edc733926e53d9d11ca5f6e949a30b85bd0eabff0b38da5591377769b WatchSource:0}: Error finding container e6cfd04edc733926e53d9d11ca5f6e949a30b85bd0eabff0b38da5591377769b: Status 404 returned error can't find the container with id e6cfd04edc733926e53d9d11ca5f6e949a30b85bd0eabff0b38da5591377769b Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.382704 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:06:11 crc kubenswrapper[4710]: E1002 13:06:11.422177 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ed1b30c_53d5_437e_b5c2_d5526668c15d.slice/crio-b48cb2a23135d3dd9b627dd5b80630befe8bf029fdd2fc9a170428e47600bc25.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdea4fdc_8bd8_4a6a_afa1_e0574adfce51.slice/crio-a793ca2ff867931392f02910af50c697e978cc00a899303389ea43e4c72f83df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ed1b30c_53d5_437e_b5c2_d5526668c15d.slice/crio-conmon-b48cb2a23135d3dd9b627dd5b80630befe8bf029fdd2fc9a170428e47600bc25.scope\": RecentStats: unable to find data in memory cache]" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.516146 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzx99\" (UniqueName: \"kubernetes.io/projected/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-kube-api-access-jzx99\") pod \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.516510 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-dns-svc\") pod \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.516607 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-config\") pod \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\" (UID: \"4d4485f3-00c0-421a-bb40-1deb2bcad3e1\") " Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.521256 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-kube-api-access-jzx99" (OuterVolumeSpecName: "kube-api-access-jzx99") pod "4d4485f3-00c0-421a-bb40-1deb2bcad3e1" (UID: "4d4485f3-00c0-421a-bb40-1deb2bcad3e1"). InnerVolumeSpecName "kube-api-access-jzx99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.564965 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-config" (OuterVolumeSpecName: "config") pod "4d4485f3-00c0-421a-bb40-1deb2bcad3e1" (UID: "4d4485f3-00c0-421a-bb40-1deb2bcad3e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.578966 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d4485f3-00c0-421a-bb40-1deb2bcad3e1" (UID: "4d4485f3-00c0-421a-bb40-1deb2bcad3e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.618874 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzx99\" (UniqueName: \"kubernetes.io/projected/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-kube-api-access-jzx99\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.618909 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.618920 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d4485f3-00c0-421a-bb40-1deb2bcad3e1-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.636573 4710 generic.go:334] "Generic (PLEG): container finished" podID="53530783-8e7c-4e44-9484-6e5e9ae70bd2" containerID="f228ccc4d1c2968d9fd2fc7f67215541470459872262a71452cec41c7c69d6ad" exitCode=0 Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.636657 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-79f7g" event={"ID":"53530783-8e7c-4e44-9484-6e5e9ae70bd2","Type":"ContainerDied","Data":"f228ccc4d1c2968d9fd2fc7f67215541470459872262a71452cec41c7c69d6ad"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.637321 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-79f7g" event={"ID":"53530783-8e7c-4e44-9484-6e5e9ae70bd2","Type":"ContainerStarted","Data":"e6cfd04edc733926e53d9d11ca5f6e949a30b85bd0eabff0b38da5591377769b"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.655688 4710 generic.go:334] "Generic (PLEG): container finished" podID="5ed1b30c-53d5-437e-b5c2-d5526668c15d" containerID="b48cb2a23135d3dd9b627dd5b80630befe8bf029fdd2fc9a170428e47600bc25" exitCode=0 Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.655759 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-msp99" event={"ID":"5ed1b30c-53d5-437e-b5c2-d5526668c15d","Type":"ContainerDied","Data":"b48cb2a23135d3dd9b627dd5b80630befe8bf029fdd2fc9a170428e47600bc25"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.655810 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-msp99" event={"ID":"5ed1b30c-53d5-437e-b5c2-d5526668c15d","Type":"ContainerStarted","Data":"ee754d252ab3473a7a9cebc927c01eafb5376a4dd2a3f9de3057ea61542a6643"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.660864 4710 generic.go:334] "Generic (PLEG): container finished" podID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerID="7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c" exitCode=0 Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.660990 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" event={"ID":"4d4485f3-00c0-421a-bb40-1deb2bcad3e1","Type":"ContainerDied","Data":"7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.661019 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.661045 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb7dbd587-vtwc9" event={"ID":"4d4485f3-00c0-421a-bb40-1deb2bcad3e1","Type":"ContainerDied","Data":"ffb1d83ec5fa9cbe7b2aad4370583e30231cef6aad77b882e351101e6868b80f"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.661071 4710 scope.go:117] "RemoveContainer" containerID="7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.662419 4710 generic.go:334] "Generic (PLEG): container finished" podID="fdea4fdc-8bd8-4a6a-afa1-e0574adfce51" containerID="a793ca2ff867931392f02910af50c697e978cc00a899303389ea43e4c72f83df" exitCode=0 Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.662500 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hzh54" event={"ID":"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51","Type":"ContainerDied","Data":"a793ca2ff867931392f02910af50c697e978cc00a899303389ea43e4c72f83df"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.662538 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hzh54" event={"ID":"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51","Type":"ContainerStarted","Data":"a334ae25fac74adec0ef4dc45fa1676fb0ac3cf7fbe6278f12eece8b1fa77145"} Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.689810 4710 scope.go:117] "RemoveContainer" containerID="d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.723408 4710 scope.go:117] "RemoveContainer" containerID="7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c" Oct 02 13:06:11 crc kubenswrapper[4710]: E1002 13:06:11.723900 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c\": container with ID starting with 7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c not found: ID does not exist" containerID="7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.723941 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c"} err="failed to get container status \"7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c\": rpc error: code = NotFound desc = could not find container \"7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c\": container with ID starting with 7e9c314ed25f4e7990936ae15e8a61ed11a924a5e94380481168f53a76bec03c not found: ID does not exist" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.723967 4710 scope.go:117] "RemoveContainer" containerID="d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11" Oct 02 13:06:11 crc kubenswrapper[4710]: E1002 13:06:11.724299 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11\": container with ID starting with d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11 not found: ID does not exist" containerID="d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.724331 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11"} err="failed to get container status \"d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11\": rpc error: code = NotFound desc = could not find container \"d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11\": container with ID starting with d70b905dc9d3f60732130328fa7eebc6feda6345a6855d791720c3d553905e11 not found: ID does not exist" Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.724408 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bb7dbd587-vtwc9"] Oct 02 13:06:11 crc kubenswrapper[4710]: I1002 13:06:11.731122 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bb7dbd587-vtwc9"] Oct 02 13:06:12 crc kubenswrapper[4710]: I1002 13:06:12.916789 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" path="/var/lib/kubelet/pods/4d4485f3-00c0-421a-bb40-1deb2bcad3e1/volumes" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.036277 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-msp99" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.145679 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66lpf\" (UniqueName: \"kubernetes.io/projected/5ed1b30c-53d5-437e-b5c2-d5526668c15d-kube-api-access-66lpf\") pod \"5ed1b30c-53d5-437e-b5c2-d5526668c15d\" (UID: \"5ed1b30c-53d5-437e-b5c2-d5526668c15d\") " Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.149491 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ed1b30c-53d5-437e-b5c2-d5526668c15d-kube-api-access-66lpf" (OuterVolumeSpecName: "kube-api-access-66lpf") pod "5ed1b30c-53d5-437e-b5c2-d5526668c15d" (UID: "5ed1b30c-53d5-437e-b5c2-d5526668c15d"). InnerVolumeSpecName "kube-api-access-66lpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.214947 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.223461 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.250420 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66lpf\" (UniqueName: \"kubernetes.io/projected/5ed1b30c-53d5-437e-b5c2-d5526668c15d-kube-api-access-66lpf\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.351385 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqm6z\" (UniqueName: \"kubernetes.io/projected/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51-kube-api-access-dqm6z\") pod \"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51\" (UID: \"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51\") " Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.351463 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cx2c\" (UniqueName: \"kubernetes.io/projected/53530783-8e7c-4e44-9484-6e5e9ae70bd2-kube-api-access-4cx2c\") pod \"53530783-8e7c-4e44-9484-6e5e9ae70bd2\" (UID: \"53530783-8e7c-4e44-9484-6e5e9ae70bd2\") " Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.354891 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51-kube-api-access-dqm6z" (OuterVolumeSpecName: "kube-api-access-dqm6z") pod "fdea4fdc-8bd8-4a6a-afa1-e0574adfce51" (UID: "fdea4fdc-8bd8-4a6a-afa1-e0574adfce51"). InnerVolumeSpecName "kube-api-access-dqm6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.367569 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53530783-8e7c-4e44-9484-6e5e9ae70bd2-kube-api-access-4cx2c" (OuterVolumeSpecName: "kube-api-access-4cx2c") pod "53530783-8e7c-4e44-9484-6e5e9ae70bd2" (UID: "53530783-8e7c-4e44-9484-6e5e9ae70bd2"). InnerVolumeSpecName "kube-api-access-4cx2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.453479 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqm6z\" (UniqueName: \"kubernetes.io/projected/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51-kube-api-access-dqm6z\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.453516 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cx2c\" (UniqueName: \"kubernetes.io/projected/53530783-8e7c-4e44-9484-6e5e9ae70bd2-kube-api-access-4cx2c\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.685170 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerStarted","Data":"7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6"} Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.688327 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-msp99" event={"ID":"5ed1b30c-53d5-437e-b5c2-d5526668c15d","Type":"ContainerDied","Data":"ee754d252ab3473a7a9cebc927c01eafb5376a4dd2a3f9de3057ea61542a6643"} Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.688372 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee754d252ab3473a7a9cebc927c01eafb5376a4dd2a3f9de3057ea61542a6643" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.688438 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-msp99" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.703026 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hzh54" event={"ID":"fdea4fdc-8bd8-4a6a-afa1-e0574adfce51","Type":"ContainerDied","Data":"a334ae25fac74adec0ef4dc45fa1676fb0ac3cf7fbe6278f12eece8b1fa77145"} Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.703069 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a334ae25fac74adec0ef4dc45fa1676fb0ac3cf7fbe6278f12eece8b1fa77145" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.703134 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hzh54" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.709315 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-79f7g" event={"ID":"53530783-8e7c-4e44-9484-6e5e9ae70bd2","Type":"ContainerDied","Data":"e6cfd04edc733926e53d9d11ca5f6e949a30b85bd0eabff0b38da5591377769b"} Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.709362 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6cfd04edc733926e53d9d11ca5f6e949a30b85bd0eabff0b38da5591377769b" Oct 02 13:06:13 crc kubenswrapper[4710]: I1002 13:06:13.709660 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-79f7g" Oct 02 13:06:14 crc kubenswrapper[4710]: I1002 13:06:14.170897 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 13:06:17 crc kubenswrapper[4710]: I1002 13:06:17.318906 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:17 crc kubenswrapper[4710]: E1002 13:06:17.319611 4710 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 13:06:17 crc kubenswrapper[4710]: E1002 13:06:17.319626 4710 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 13:06:17 crc kubenswrapper[4710]: E1002 13:06:17.319675 4710 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift podName:67031d3a-15f2-4b86-92d7-b525a3a8420e nodeName:}" failed. No retries permitted until 2025-10-02 13:06:33.319659939 +0000 UTC m=+1037.426070822 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift") pod "swift-storage-0" (UID: "67031d3a-15f2-4b86-92d7-b525a3a8420e") : configmap "swift-ring-files" not found Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.150468 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c3fa-account-create-ml5qg"] Oct 02 13:06:18 crc kubenswrapper[4710]: E1002 13:06:18.153139 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53530783-8e7c-4e44-9484-6e5e9ae70bd2" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153185 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="53530783-8e7c-4e44-9484-6e5e9ae70bd2" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: E1002 13:06:18.153203 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdc93b0-6465-43e9-8d1f-045f2ee24460" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153211 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdc93b0-6465-43e9-8d1f-045f2ee24460" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: E1002 13:06:18.153229 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerName="init" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153240 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerName="init" Oct 02 13:06:18 crc kubenswrapper[4710]: E1002 13:06:18.153264 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerName="dnsmasq-dns" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153272 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerName="dnsmasq-dns" Oct 02 13:06:18 crc kubenswrapper[4710]: E1002 13:06:18.153281 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdea4fdc-8bd8-4a6a-afa1-e0574adfce51" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153288 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdea4fdc-8bd8-4a6a-afa1-e0574adfce51" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: E1002 13:06:18.153301 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed1b30c-53d5-437e-b5c2-d5526668c15d" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153309 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed1b30c-53d5-437e-b5c2-d5526668c15d" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153611 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdc93b0-6465-43e9-8d1f-045f2ee24460" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153647 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4485f3-00c0-421a-bb40-1deb2bcad3e1" containerName="dnsmasq-dns" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153664 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ed1b30c-53d5-437e-b5c2-d5526668c15d" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153678 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdea4fdc-8bd8-4a6a-afa1-e0574adfce51" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.153688 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="53530783-8e7c-4e44-9484-6e5e9ae70bd2" containerName="mariadb-database-create" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.154491 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.158389 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.161142 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c3fa-account-create-ml5qg"] Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.340382 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7d2n\" (UniqueName: \"kubernetes.io/projected/d1f96697-8eea-4ade-a5f4-a51986a0f83f-kube-api-access-g7d2n\") pod \"keystone-c3fa-account-create-ml5qg\" (UID: \"d1f96697-8eea-4ade-a5f4-a51986a0f83f\") " pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.444009 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7d2n\" (UniqueName: \"kubernetes.io/projected/d1f96697-8eea-4ade-a5f4-a51986a0f83f-kube-api-access-g7d2n\") pod \"keystone-c3fa-account-create-ml5qg\" (UID: \"d1f96697-8eea-4ade-a5f4-a51986a0f83f\") " pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.467640 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7d2n\" (UniqueName: \"kubernetes.io/projected/d1f96697-8eea-4ade-a5f4-a51986a0f83f-kube-api-access-g7d2n\") pod \"keystone-c3fa-account-create-ml5qg\" (UID: \"d1f96697-8eea-4ade-a5f4-a51986a0f83f\") " pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:18 crc kubenswrapper[4710]: I1002 13:06:18.487154 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.147083 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c3fa-account-create-ml5qg"] Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.761791 4710 generic.go:334] "Generic (PLEG): container finished" podID="989f6404-4404-4e77-b3bc-f3f27a7e7520" containerID="eedb51a6b5275fb5e32aafb1717087a039bd72fd2c3e04367920aa94e4fb032e" exitCode=0 Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.761917 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xvqwj" event={"ID":"989f6404-4404-4e77-b3bc-f3f27a7e7520","Type":"ContainerDied","Data":"eedb51a6b5275fb5e32aafb1717087a039bd72fd2c3e04367920aa94e4fb032e"} Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.765082 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerStarted","Data":"896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768"} Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.767805 4710 generic.go:334] "Generic (PLEG): container finished" podID="d1f96697-8eea-4ade-a5f4-a51986a0f83f" containerID="85a23d9e67017a47c214124a314e55456698490a9ce07b92aa893470b2979bbb" exitCode=0 Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.767855 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c3fa-account-create-ml5qg" event={"ID":"d1f96697-8eea-4ade-a5f4-a51986a0f83f","Type":"ContainerDied","Data":"85a23d9e67017a47c214124a314e55456698490a9ce07b92aa893470b2979bbb"} Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.767882 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c3fa-account-create-ml5qg" event={"ID":"d1f96697-8eea-4ade-a5f4-a51986a0f83f","Type":"ContainerStarted","Data":"6393720f40f3db7cf46087b7d243cd420d5509c4c01f904effb36b760561028d"} Oct 02 13:06:19 crc kubenswrapper[4710]: I1002 13:06:19.811140 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=15.896659053 podStartE2EDuration="49.811121298s" podCreationTimestamp="2025-10-02 13:05:30 +0000 UTC" firstStartedPulling="2025-10-02 13:05:44.998920417 +0000 UTC m=+989.105331300" lastFinishedPulling="2025-10-02 13:06:18.913382652 +0000 UTC m=+1023.019793545" observedRunningTime="2025-10-02 13:06:19.802953684 +0000 UTC m=+1023.909364567" watchObservedRunningTime="2025-10-02 13:06:19.811121298 +0000 UTC m=+1023.917532191" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.272316 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-7e6f-account-create-8tfx4"] Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.273389 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.275941 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.289722 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-7e6f-account-create-8tfx4"] Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.378326 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk4b7\" (UniqueName: \"kubernetes.io/projected/882558ba-032b-4377-a806-ae3760562289-kube-api-access-xk4b7\") pod \"watcher-7e6f-account-create-8tfx4\" (UID: \"882558ba-032b-4377-a806-ae3760562289\") " pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.479928 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk4b7\" (UniqueName: \"kubernetes.io/projected/882558ba-032b-4377-a806-ae3760562289-kube-api-access-xk4b7\") pod \"watcher-7e6f-account-create-8tfx4\" (UID: \"882558ba-032b-4377-a806-ae3760562289\") " pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.504841 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk4b7\" (UniqueName: \"kubernetes.io/projected/882558ba-032b-4377-a806-ae3760562289-kube-api-access-xk4b7\") pod \"watcher-7e6f-account-create-8tfx4\" (UID: \"882558ba-032b-4377-a806-ae3760562289\") " pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.595994 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.778557 4710 generic.go:334] "Generic (PLEG): container finished" podID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerID="6b868c2abc568ca5d3358f49917fba3e3214d607ef5d3ce34983a6c316e15ab3" exitCode=0 Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.778621 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"770af87f-9638-4cdb-a10d-2b40d50302ca","Type":"ContainerDied","Data":"6b868c2abc568ca5d3358f49917fba3e3214d607ef5d3ce34983a6c316e15ab3"} Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.782395 4710 generic.go:334] "Generic (PLEG): container finished" podID="98c2f825-d5c6-47e7-a96d-8cf7227b0a24" containerID="11d9d62f9794dafce40bfec1992b6e85df6f088b226022fb0e13a8880b0312d7" exitCode=0 Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.782483 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"98c2f825-d5c6-47e7-a96d-8cf7227b0a24","Type":"ContainerDied","Data":"11d9d62f9794dafce40bfec1992b6e85df6f088b226022fb0e13a8880b0312d7"} Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.792514 4710 generic.go:334] "Generic (PLEG): container finished" podID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerID="6987be090ed6cf0c897ec0e533b9d01229b7de9b3109dc7f672c9f5516cfef82" exitCode=0 Oct 02 13:06:20 crc kubenswrapper[4710]: I1002 13:06:20.792624 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7fb5713-6a35-4676-b466-6a43a4037f7b","Type":"ContainerDied","Data":"6987be090ed6cf0c897ec0e533b9d01229b7de9b3109dc7f672c9f5516cfef82"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.121171 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-7e6f-account-create-8tfx4"] Oct 02 13:06:21 crc kubenswrapper[4710]: W1002 13:06:21.136776 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod882558ba_032b_4377_a806_ae3760562289.slice/crio-9a00611a09b3a701aecd2508c8257315289aeb96c642bcc4526b66a267170ae4 WatchSource:0}: Error finding container 9a00611a09b3a701aecd2508c8257315289aeb96c642bcc4526b66a267170ae4: Status 404 returned error can't find the container with id 9a00611a09b3a701aecd2508c8257315289aeb96c642bcc4526b66a267170ae4 Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.209368 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.253940 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.294207 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-dispersionconf\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.294342 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-swiftconf\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.294373 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-combined-ca-bundle\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.295435 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-ring-data-devices\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.295480 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghtj7\" (UniqueName: \"kubernetes.io/projected/989f6404-4404-4e77-b3bc-f3f27a7e7520-kube-api-access-ghtj7\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.295572 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-scripts\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.295638 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/989f6404-4404-4e77-b3bc-f3f27a7e7520-etc-swift\") pod \"989f6404-4404-4e77-b3bc-f3f27a7e7520\" (UID: \"989f6404-4404-4e77-b3bc-f3f27a7e7520\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.296306 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.297446 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/989f6404-4404-4e77-b3bc-f3f27a7e7520-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.307305 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.309374 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989f6404-4404-4e77-b3bc-f3f27a7e7520-kube-api-access-ghtj7" (OuterVolumeSpecName: "kube-api-access-ghtj7") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "kube-api-access-ghtj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.322428 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-scripts" (OuterVolumeSpecName: "scripts") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.331807 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.332418 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "989f6404-4404-4e77-b3bc-f3f27a7e7520" (UID: "989f6404-4404-4e77-b3bc-f3f27a7e7520"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.390601 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.397592 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7d2n\" (UniqueName: \"kubernetes.io/projected/d1f96697-8eea-4ade-a5f4-a51986a0f83f-kube-api-access-g7d2n\") pod \"d1f96697-8eea-4ade-a5f4-a51986a0f83f\" (UID: \"d1f96697-8eea-4ade-a5f4-a51986a0f83f\") " Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398011 4710 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/989f6404-4404-4e77-b3bc-f3f27a7e7520-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398037 4710 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398049 4710 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398085 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989f6404-4404-4e77-b3bc-f3f27a7e7520-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398100 4710 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398110 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghtj7\" (UniqueName: \"kubernetes.io/projected/989f6404-4404-4e77-b3bc-f3f27a7e7520-kube-api-access-ghtj7\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.398122 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/989f6404-4404-4e77-b3bc-f3f27a7e7520-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.402428 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f96697-8eea-4ade-a5f4-a51986a0f83f-kube-api-access-g7d2n" (OuterVolumeSpecName: "kube-api-access-g7d2n") pod "d1f96697-8eea-4ade-a5f4-a51986a0f83f" (UID: "d1f96697-8eea-4ade-a5f4-a51986a0f83f"). InnerVolumeSpecName "kube-api-access-g7d2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.499307 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7d2n\" (UniqueName: \"kubernetes.io/projected/d1f96697-8eea-4ade-a5f4-a51986a0f83f-kube-api-access-g7d2n\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.801965 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"98c2f825-d5c6-47e7-a96d-8cf7227b0a24","Type":"ContainerStarted","Data":"bb1141cd6f01e789082dd35a70cf392fdf1f58aad1ce24504b6fda93cfcdb201"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.804186 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.804641 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7fb5713-6a35-4676-b466-6a43a4037f7b","Type":"ContainerStarted","Data":"5e0e2863df8edcbcd5a8ff5b36093c5369fa9833fab9d8d8b15a4dda623e23e3"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.804785 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.806769 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"770af87f-9638-4cdb-a10d-2b40d50302ca","Type":"ContainerStarted","Data":"2a53df3200ea0655a7b8d77a58300f314704d1f704aac912604d3e7e26ffdf3e"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.807272 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.808685 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xvqwj" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.808667 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xvqwj" event={"ID":"989f6404-4404-4e77-b3bc-f3f27a7e7520","Type":"ContainerDied","Data":"87f25bcc9dccbb2ab7701ca64b9aa64f4bc41a28e8644155d8562a41088cea20"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.809044 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87f25bcc9dccbb2ab7701ca64b9aa64f4bc41a28e8644155d8562a41088cea20" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.810297 4710 generic.go:334] "Generic (PLEG): container finished" podID="882558ba-032b-4377-a806-ae3760562289" containerID="19f69cc552f88b7ed533dc2618f92ba1e0d58eb7ae0f5fd9330bbc19a0738005" exitCode=0 Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.810428 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7e6f-account-create-8tfx4" event={"ID":"882558ba-032b-4377-a806-ae3760562289","Type":"ContainerDied","Data":"19f69cc552f88b7ed533dc2618f92ba1e0d58eb7ae0f5fd9330bbc19a0738005"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.810544 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7e6f-account-create-8tfx4" event={"ID":"882558ba-032b-4377-a806-ae3760562289","Type":"ContainerStarted","Data":"9a00611a09b3a701aecd2508c8257315289aeb96c642bcc4526b66a267170ae4"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.812360 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c3fa-account-create-ml5qg" event={"ID":"d1f96697-8eea-4ade-a5f4-a51986a0f83f","Type":"ContainerDied","Data":"6393720f40f3db7cf46087b7d243cd420d5509c4c01f904effb36b760561028d"} Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.812396 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6393720f40f3db7cf46087b7d243cd420d5509c4c01f904effb36b760561028d" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.812367 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c3fa-account-create-ml5qg" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.868061 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=54.532859913 podStartE2EDuration="58.868038498s" podCreationTimestamp="2025-10-02 13:05:23 +0000 UTC" firstStartedPulling="2025-10-02 13:05:42.235128108 +0000 UTC m=+986.341538991" lastFinishedPulling="2025-10-02 13:05:46.570306693 +0000 UTC m=+990.676717576" observedRunningTime="2025-10-02 13:06:21.854523061 +0000 UTC m=+1025.960933964" watchObservedRunningTime="2025-10-02 13:06:21.868038498 +0000 UTC m=+1025.974449381" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.910646 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=53.733579006 podStartE2EDuration="58.910630832s" podCreationTimestamp="2025-10-02 13:05:23 +0000 UTC" firstStartedPulling="2025-10-02 13:05:41.39099833 +0000 UTC m=+985.497409213" lastFinishedPulling="2025-10-02 13:05:46.568050156 +0000 UTC m=+990.674461039" observedRunningTime="2025-10-02 13:06:21.904630152 +0000 UTC m=+1026.011041045" watchObservedRunningTime="2025-10-02 13:06:21.910630832 +0000 UTC m=+1026.017041715" Oct 02 13:06:21 crc kubenswrapper[4710]: I1002 13:06:21.933037 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=57.240455079 podStartE2EDuration="58.933018191s" podCreationTimestamp="2025-10-02 13:05:23 +0000 UTC" firstStartedPulling="2025-10-02 13:05:44.997134032 +0000 UTC m=+989.103544915" lastFinishedPulling="2025-10-02 13:05:46.689697144 +0000 UTC m=+990.796108027" observedRunningTime="2025-10-02 13:06:21.926195761 +0000 UTC m=+1026.032606654" watchObservedRunningTime="2025-10-02 13:06:21.933018191 +0000 UTC m=+1026.039429074" Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.530103 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.530435 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.530490 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.531141 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be59f3d65c42b4df4c251f325b3eef5d2ce156af0437d5494544eb88ffa0a1fd"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.531210 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://be59f3d65c42b4df4c251f325b3eef5d2ce156af0437d5494544eb88ffa0a1fd" gracePeriod=600 Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.822981 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="be59f3d65c42b4df4c251f325b3eef5d2ce156af0437d5494544eb88ffa0a1fd" exitCode=0 Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.823093 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"be59f3d65c42b4df4c251f325b3eef5d2ce156af0437d5494544eb88ffa0a1fd"} Oct 02 13:06:22 crc kubenswrapper[4710]: I1002 13:06:22.823157 4710 scope.go:117] "RemoveContainer" containerID="2554f6884f6fea884aeb14ab54b4450d5c8dcfe9721cc87320ad6736f5b0d86e" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.163152 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.227049 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk4b7\" (UniqueName: \"kubernetes.io/projected/882558ba-032b-4377-a806-ae3760562289-kube-api-access-xk4b7\") pod \"882558ba-032b-4377-a806-ae3760562289\" (UID: \"882558ba-032b-4377-a806-ae3760562289\") " Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.233031 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/882558ba-032b-4377-a806-ae3760562289-kube-api-access-xk4b7" (OuterVolumeSpecName: "kube-api-access-xk4b7") pod "882558ba-032b-4377-a806-ae3760562289" (UID: "882558ba-032b-4377-a806-ae3760562289"). InnerVolumeSpecName "kube-api-access-xk4b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.329083 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk4b7\" (UniqueName: \"kubernetes.io/projected/882558ba-032b-4377-a806-ae3760562289-kube-api-access-xk4b7\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.733221 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fdf1-account-create-9z6r2"] Oct 02 13:06:23 crc kubenswrapper[4710]: E1002 13:06:23.733972 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f96697-8eea-4ade-a5f4-a51986a0f83f" containerName="mariadb-account-create" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.734076 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f96697-8eea-4ade-a5f4-a51986a0f83f" containerName="mariadb-account-create" Oct 02 13:06:23 crc kubenswrapper[4710]: E1002 13:06:23.734170 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882558ba-032b-4377-a806-ae3760562289" containerName="mariadb-account-create" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.734248 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="882558ba-032b-4377-a806-ae3760562289" containerName="mariadb-account-create" Oct 02 13:06:23 crc kubenswrapper[4710]: E1002 13:06:23.734336 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989f6404-4404-4e77-b3bc-f3f27a7e7520" containerName="swift-ring-rebalance" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.734411 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="989f6404-4404-4e77-b3bc-f3f27a7e7520" containerName="swift-ring-rebalance" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.734683 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="882558ba-032b-4377-a806-ae3760562289" containerName="mariadb-account-create" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.734879 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f96697-8eea-4ade-a5f4-a51986a0f83f" containerName="mariadb-account-create" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.734982 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="989f6404-4404-4e77-b3bc-f3f27a7e7520" containerName="swift-ring-rebalance" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.735801 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.739183 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.747956 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fdf1-account-create-9z6r2"] Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.834493 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"51e94e7f2d48ee2b42323dae0e2a0810cf3081a0dd80ed3b534b22c393fd779b"} Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.837659 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7mpd\" (UniqueName: \"kubernetes.io/projected/dd700f13-8732-4f7f-bbaf-ba5f8156ba97-kube-api-access-x7mpd\") pod \"glance-fdf1-account-create-9z6r2\" (UID: \"dd700f13-8732-4f7f-bbaf-ba5f8156ba97\") " pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.838116 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7e6f-account-create-8tfx4" event={"ID":"882558ba-032b-4377-a806-ae3760562289","Type":"ContainerDied","Data":"9a00611a09b3a701aecd2508c8257315289aeb96c642bcc4526b66a267170ae4"} Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.838149 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a00611a09b3a701aecd2508c8257315289aeb96c642bcc4526b66a267170ae4" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.838202 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7e6f-account-create-8tfx4" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.939684 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7mpd\" (UniqueName: \"kubernetes.io/projected/dd700f13-8732-4f7f-bbaf-ba5f8156ba97-kube-api-access-x7mpd\") pod \"glance-fdf1-account-create-9z6r2\" (UID: \"dd700f13-8732-4f7f-bbaf-ba5f8156ba97\") " pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:23 crc kubenswrapper[4710]: I1002 13:06:23.965488 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7mpd\" (UniqueName: \"kubernetes.io/projected/dd700f13-8732-4f7f-bbaf-ba5f8156ba97-kube-api-access-x7mpd\") pod \"glance-fdf1-account-create-9z6r2\" (UID: \"dd700f13-8732-4f7f-bbaf-ba5f8156ba97\") " pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:24 crc kubenswrapper[4710]: I1002 13:06:24.072725 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:24 crc kubenswrapper[4710]: I1002 13:06:24.537719 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fdf1-account-create-9z6r2"] Oct 02 13:06:24 crc kubenswrapper[4710]: I1002 13:06:24.848068 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fdf1-account-create-9z6r2" event={"ID":"dd700f13-8732-4f7f-bbaf-ba5f8156ba97","Type":"ContainerStarted","Data":"dafafd991dee79759d06a6a546daf02dce167312d311979e539fed3a6e71c841"} Oct 02 13:06:24 crc kubenswrapper[4710]: I1002 13:06:24.848118 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fdf1-account-create-9z6r2" event={"ID":"dd700f13-8732-4f7f-bbaf-ba5f8156ba97","Type":"ContainerStarted","Data":"e576fca73ea6d44e0714a2728d6ecfc4b7df78928b4e2a29980648b4e7c8c7c5"} Oct 02 13:06:24 crc kubenswrapper[4710]: I1002 13:06:24.870805 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-fdf1-account-create-9z6r2" podStartSLOduration=1.8707884940000001 podStartE2EDuration="1.870788494s" podCreationTimestamp="2025-10-02 13:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:06:24.862110748 +0000 UTC m=+1028.968521651" watchObservedRunningTime="2025-10-02 13:06:24.870788494 +0000 UTC m=+1028.977199367" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.321008 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fp7wl" podUID="d2d9ea68-a957-4056-867d-18e0535bec60" containerName="ovn-controller" probeResult="failure" output=< Oct 02 13:06:25 crc kubenswrapper[4710]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 13:06:25 crc kubenswrapper[4710]: > Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.351861 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.353168 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jd2lt" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.591231 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fp7wl-config-csqx7"] Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.592241 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.594614 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.609186 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fp7wl-config-csqx7"] Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.664107 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-additional-scripts\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.664152 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-scripts\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.664189 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.664363 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjvgz\" (UniqueName: \"kubernetes.io/projected/66f69417-b5bc-497a-b8b0-ae14e86b3f49-kube-api-access-xjvgz\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.664464 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-log-ovn\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.664605 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run-ovn\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.766522 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.766603 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjvgz\" (UniqueName: \"kubernetes.io/projected/66f69417-b5bc-497a-b8b0-ae14e86b3f49-kube-api-access-xjvgz\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.766662 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-log-ovn\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.766760 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run-ovn\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.766818 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-additional-scripts\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.766842 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-scripts\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.767325 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.767345 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run-ovn\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.767510 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-log-ovn\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.767988 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-additional-scripts\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.769066 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-scripts\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.785665 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjvgz\" (UniqueName: \"kubernetes.io/projected/66f69417-b5bc-497a-b8b0-ae14e86b3f49-kube-api-access-xjvgz\") pod \"ovn-controller-fp7wl-config-csqx7\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.856917 4710 generic.go:334] "Generic (PLEG): container finished" podID="dd700f13-8732-4f7f-bbaf-ba5f8156ba97" containerID="dafafd991dee79759d06a6a546daf02dce167312d311979e539fed3a6e71c841" exitCode=0 Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.857115 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fdf1-account-create-9z6r2" event={"ID":"dd700f13-8732-4f7f-bbaf-ba5f8156ba97","Type":"ContainerDied","Data":"dafafd991dee79759d06a6a546daf02dce167312d311979e539fed3a6e71c841"} Oct 02 13:06:25 crc kubenswrapper[4710]: I1002 13:06:25.910760 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:26 crc kubenswrapper[4710]: I1002 13:06:26.353279 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fp7wl-config-csqx7"] Oct 02 13:06:26 crc kubenswrapper[4710]: I1002 13:06:26.865781 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-csqx7" event={"ID":"66f69417-b5bc-497a-b8b0-ae14e86b3f49","Type":"ContainerStarted","Data":"fdee93596cd8050671b0eb9eb2f1293b5198c88057e712355864ad2002099d9f"} Oct 02 13:06:26 crc kubenswrapper[4710]: I1002 13:06:26.866039 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-csqx7" event={"ID":"66f69417-b5bc-497a-b8b0-ae14e86b3f49","Type":"ContainerStarted","Data":"7af8a0ef21955fb9ac09b0ae2676a92888d48fa73edd614fcd16b14271ddb1a6"} Oct 02 13:06:26 crc kubenswrapper[4710]: I1002 13:06:26.890708 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fp7wl-config-csqx7" podStartSLOduration=1.890689199 podStartE2EDuration="1.890689199s" podCreationTimestamp="2025-10-02 13:06:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:06:26.884646818 +0000 UTC m=+1030.991057701" watchObservedRunningTime="2025-10-02 13:06:26.890689199 +0000 UTC m=+1030.997100082" Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.257277 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.296381 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7mpd\" (UniqueName: \"kubernetes.io/projected/dd700f13-8732-4f7f-bbaf-ba5f8156ba97-kube-api-access-x7mpd\") pod \"dd700f13-8732-4f7f-bbaf-ba5f8156ba97\" (UID: \"dd700f13-8732-4f7f-bbaf-ba5f8156ba97\") " Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.302626 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd700f13-8732-4f7f-bbaf-ba5f8156ba97-kube-api-access-x7mpd" (OuterVolumeSpecName: "kube-api-access-x7mpd") pod "dd700f13-8732-4f7f-bbaf-ba5f8156ba97" (UID: "dd700f13-8732-4f7f-bbaf-ba5f8156ba97"). InnerVolumeSpecName "kube-api-access-x7mpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.398258 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7mpd\" (UniqueName: \"kubernetes.io/projected/dd700f13-8732-4f7f-bbaf-ba5f8156ba97-kube-api-access-x7mpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.873687 4710 generic.go:334] "Generic (PLEG): container finished" podID="66f69417-b5bc-497a-b8b0-ae14e86b3f49" containerID="fdee93596cd8050671b0eb9eb2f1293b5198c88057e712355864ad2002099d9f" exitCode=0 Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.873753 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-csqx7" event={"ID":"66f69417-b5bc-497a-b8b0-ae14e86b3f49","Type":"ContainerDied","Data":"fdee93596cd8050671b0eb9eb2f1293b5198c88057e712355864ad2002099d9f"} Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.875558 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fdf1-account-create-9z6r2" event={"ID":"dd700f13-8732-4f7f-bbaf-ba5f8156ba97","Type":"ContainerDied","Data":"e576fca73ea6d44e0714a2728d6ecfc4b7df78928b4e2a29980648b4e7c8c7c5"} Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.875587 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e576fca73ea6d44e0714a2728d6ecfc4b7df78928b4e2a29980648b4e7c8c7c5" Oct 02 13:06:27 crc kubenswrapper[4710]: I1002 13:06:27.875629 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fdf1-account-create-9z6r2" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.454224 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d43-account-create-d6l9p"] Oct 02 13:06:28 crc kubenswrapper[4710]: E1002 13:06:28.454637 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd700f13-8732-4f7f-bbaf-ba5f8156ba97" containerName="mariadb-account-create" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.454659 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd700f13-8732-4f7f-bbaf-ba5f8156ba97" containerName="mariadb-account-create" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.454871 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd700f13-8732-4f7f-bbaf-ba5f8156ba97" containerName="mariadb-account-create" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.455579 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.457820 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.493696 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d43-account-create-d6l9p"] Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.515758 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85hnz\" (UniqueName: \"kubernetes.io/projected/ce991bb5-1e98-40f5-8395-b1997433b236-kube-api-access-85hnz\") pod \"placement-7d43-account-create-d6l9p\" (UID: \"ce991bb5-1e98-40f5-8395-b1997433b236\") " pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.617038 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85hnz\" (UniqueName: \"kubernetes.io/projected/ce991bb5-1e98-40f5-8395-b1997433b236-kube-api-access-85hnz\") pod \"placement-7d43-account-create-d6l9p\" (UID: \"ce991bb5-1e98-40f5-8395-b1997433b236\") " pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.640266 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85hnz\" (UniqueName: \"kubernetes.io/projected/ce991bb5-1e98-40f5-8395-b1997433b236-kube-api-access-85hnz\") pod \"placement-7d43-account-create-d6l9p\" (UID: \"ce991bb5-1e98-40f5-8395-b1997433b236\") " pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.771646 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.880285 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-6htbd"] Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.881734 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.885861 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.886075 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6fkzg" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.919517 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6htbd"] Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.921445 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-db-sync-config-data\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.921490 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2w26\" (UniqueName: \"kubernetes.io/projected/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-kube-api-access-m2w26\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.921566 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-config-data\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:28 crc kubenswrapper[4710]: I1002 13:06:28.921621 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-combined-ca-bundle\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.023012 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2w26\" (UniqueName: \"kubernetes.io/projected/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-kube-api-access-m2w26\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.023118 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-config-data\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.023154 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-combined-ca-bundle\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.023225 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-db-sync-config-data\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.033615 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-combined-ca-bundle\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.033670 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-config-data\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.035795 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-db-sync-config-data\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.042178 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2w26\" (UniqueName: \"kubernetes.io/projected/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-kube-api-access-m2w26\") pod \"glance-db-sync-6htbd\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.214079 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6htbd" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.254817 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.295419 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d43-account-create-d6l9p"] Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.326327 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run\") pod \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.326387 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run-ovn\") pod \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.326421 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-scripts\") pod \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.326496 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-log-ovn\") pod \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.326537 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-additional-scripts\") pod \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.326565 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjvgz\" (UniqueName: \"kubernetes.io/projected/66f69417-b5bc-497a-b8b0-ae14e86b3f49-kube-api-access-xjvgz\") pod \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\" (UID: \"66f69417-b5bc-497a-b8b0-ae14e86b3f49\") " Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.327057 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run" (OuterVolumeSpecName: "var-run") pod "66f69417-b5bc-497a-b8b0-ae14e86b3f49" (UID: "66f69417-b5bc-497a-b8b0-ae14e86b3f49"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.327102 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "66f69417-b5bc-497a-b8b0-ae14e86b3f49" (UID: "66f69417-b5bc-497a-b8b0-ae14e86b3f49"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.328133 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-scripts" (OuterVolumeSpecName: "scripts") pod "66f69417-b5bc-497a-b8b0-ae14e86b3f49" (UID: "66f69417-b5bc-497a-b8b0-ae14e86b3f49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.328168 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "66f69417-b5bc-497a-b8b0-ae14e86b3f49" (UID: "66f69417-b5bc-497a-b8b0-ae14e86b3f49"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.329004 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "66f69417-b5bc-497a-b8b0-ae14e86b3f49" (UID: "66f69417-b5bc-497a-b8b0-ae14e86b3f49"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.348882 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f69417-b5bc-497a-b8b0-ae14e86b3f49-kube-api-access-xjvgz" (OuterVolumeSpecName: "kube-api-access-xjvgz") pod "66f69417-b5bc-497a-b8b0-ae14e86b3f49" (UID: "66f69417-b5bc-497a-b8b0-ae14e86b3f49"). InnerVolumeSpecName "kube-api-access-xjvgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.428971 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjvgz\" (UniqueName: \"kubernetes.io/projected/66f69417-b5bc-497a-b8b0-ae14e86b3f49-kube-api-access-xjvgz\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.429001 4710 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.429013 4710 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.429022 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.429031 4710 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66f69417-b5bc-497a-b8b0-ae14e86b3f49-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.429038 4710 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/66f69417-b5bc-497a-b8b0-ae14e86b3f49-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.806046 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6htbd"] Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.898629 4710 generic.go:334] "Generic (PLEG): container finished" podID="ce991bb5-1e98-40f5-8395-b1997433b236" containerID="4712bbc24352470c7207735dbfbd436fc4b312e6d98f93bf18c31b64577d6747" exitCode=0 Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.898683 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d43-account-create-d6l9p" event={"ID":"ce991bb5-1e98-40f5-8395-b1997433b236","Type":"ContainerDied","Data":"4712bbc24352470c7207735dbfbd436fc4b312e6d98f93bf18c31b64577d6747"} Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.899031 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d43-account-create-d6l9p" event={"ID":"ce991bb5-1e98-40f5-8395-b1997433b236","Type":"ContainerStarted","Data":"80ea21bd50bd1346884157535320364ab4744993974ebfdae0379b7b2ab794e2"} Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.900380 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6htbd" event={"ID":"52ce5f44-dff4-4fe5-aa15-e76e352b9d88","Type":"ContainerStarted","Data":"94ff6ca2311527d537d0ef953dcad1588e42d109c3e7e0b702884907cc5d0a75"} Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.902346 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-csqx7" event={"ID":"66f69417-b5bc-497a-b8b0-ae14e86b3f49","Type":"ContainerDied","Data":"7af8a0ef21955fb9ac09b0ae2676a92888d48fa73edd614fcd16b14271ddb1a6"} Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.902378 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7af8a0ef21955fb9ac09b0ae2676a92888d48fa73edd614fcd16b14271ddb1a6" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.902388 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-csqx7" Oct 02 13:06:29 crc kubenswrapper[4710]: I1002 13:06:29.989712 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fp7wl-config-csqx7"] Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.002448 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fp7wl-config-csqx7"] Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.111422 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fp7wl-config-wq2tg"] Oct 02 13:06:30 crc kubenswrapper[4710]: E1002 13:06:30.111848 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f69417-b5bc-497a-b8b0-ae14e86b3f49" containerName="ovn-config" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.111878 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f69417-b5bc-497a-b8b0-ae14e86b3f49" containerName="ovn-config" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.112072 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f69417-b5bc-497a-b8b0-ae14e86b3f49" containerName="ovn-config" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.112658 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.116932 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.128990 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fp7wl-config-wq2tg"] Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.138984 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.139034 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run-ovn\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.139102 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-additional-scripts\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.139151 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww9ms\" (UniqueName: \"kubernetes.io/projected/46749e91-fd6f-4dcd-8c10-c11352de9138-kube-api-access-ww9ms\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.139212 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-log-ovn\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.139273 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-scripts\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240369 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-additional-scripts\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240422 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww9ms\" (UniqueName: \"kubernetes.io/projected/46749e91-fd6f-4dcd-8c10-c11352de9138-kube-api-access-ww9ms\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240469 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-log-ovn\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240504 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-scripts\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240553 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240568 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run-ovn\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.240841 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run-ovn\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.241406 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-additional-scripts\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.241666 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-log-ovn\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.241900 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.243370 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-scripts\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.263070 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww9ms\" (UniqueName: \"kubernetes.io/projected/46749e91-fd6f-4dcd-8c10-c11352de9138-kube-api-access-ww9ms\") pod \"ovn-controller-fp7wl-config-wq2tg\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.332923 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-fp7wl" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.428936 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.877094 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fp7wl-config-wq2tg"] Oct 02 13:06:30 crc kubenswrapper[4710]: W1002 13:06:30.889341 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46749e91_fd6f_4dcd_8c10_c11352de9138.slice/crio-979762defae71800b9a87edabc5118677239dd47e25fc1e7e19f6bbb133f37bf WatchSource:0}: Error finding container 979762defae71800b9a87edabc5118677239dd47e25fc1e7e19f6bbb133f37bf: Status 404 returned error can't find the container with id 979762defae71800b9a87edabc5118677239dd47e25fc1e7e19f6bbb133f37bf Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.922879 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66f69417-b5bc-497a-b8b0-ae14e86b3f49" path="/var/lib/kubelet/pods/66f69417-b5bc-497a-b8b0-ae14e86b3f49/volumes" Oct 02 13:06:30 crc kubenswrapper[4710]: I1002 13:06:30.925248 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-wq2tg" event={"ID":"46749e91-fd6f-4dcd-8c10-c11352de9138","Type":"ContainerStarted","Data":"979762defae71800b9a87edabc5118677239dd47e25fc1e7e19f6bbb133f37bf"} Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.293959 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.361376 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85hnz\" (UniqueName: \"kubernetes.io/projected/ce991bb5-1e98-40f5-8395-b1997433b236-kube-api-access-85hnz\") pod \"ce991bb5-1e98-40f5-8395-b1997433b236\" (UID: \"ce991bb5-1e98-40f5-8395-b1997433b236\") " Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.377079 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce991bb5-1e98-40f5-8395-b1997433b236-kube-api-access-85hnz" (OuterVolumeSpecName: "kube-api-access-85hnz") pod "ce991bb5-1e98-40f5-8395-b1997433b236" (UID: "ce991bb5-1e98-40f5-8395-b1997433b236"). InnerVolumeSpecName "kube-api-access-85hnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.390092 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.393258 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.464902 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85hnz\" (UniqueName: \"kubernetes.io/projected/ce991bb5-1e98-40f5-8395-b1997433b236-kube-api-access-85hnz\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.933445 4710 generic.go:334] "Generic (PLEG): container finished" podID="46749e91-fd6f-4dcd-8c10-c11352de9138" containerID="d09fdff87746597e8783db624850e59db4ddbebcd80219d59a2c98e7480230a2" exitCode=0 Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.934094 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-wq2tg" event={"ID":"46749e91-fd6f-4dcd-8c10-c11352de9138","Type":"ContainerDied","Data":"d09fdff87746597e8783db624850e59db4ddbebcd80219d59a2c98e7480230a2"} Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.936518 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d43-account-create-d6l9p" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.939353 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d43-account-create-d6l9p" event={"ID":"ce991bb5-1e98-40f5-8395-b1997433b236","Type":"ContainerDied","Data":"80ea21bd50bd1346884157535320364ab4744993974ebfdae0379b7b2ab794e2"} Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.939421 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80ea21bd50bd1346884157535320364ab4744993974ebfdae0379b7b2ab794e2" Oct 02 13:06:31 crc kubenswrapper[4710]: I1002 13:06:31.940096 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.321697 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.398094 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.406652 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/67031d3a-15f2-4b86-92d7-b525a3a8420e-etc-swift\") pod \"swift-storage-0\" (UID: \"67031d3a-15f2-4b86-92d7-b525a3a8420e\") " pod="openstack/swift-storage-0" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499674 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run-ovn\") pod \"46749e91-fd6f-4dcd-8c10-c11352de9138\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499804 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "46749e91-fd6f-4dcd-8c10-c11352de9138" (UID: "46749e91-fd6f-4dcd-8c10-c11352de9138"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499817 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww9ms\" (UniqueName: \"kubernetes.io/projected/46749e91-fd6f-4dcd-8c10-c11352de9138-kube-api-access-ww9ms\") pod \"46749e91-fd6f-4dcd-8c10-c11352de9138\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499842 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-log-ovn\") pod \"46749e91-fd6f-4dcd-8c10-c11352de9138\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499872 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-scripts\") pod \"46749e91-fd6f-4dcd-8c10-c11352de9138\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499892 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run\") pod \"46749e91-fd6f-4dcd-8c10-c11352de9138\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.499917 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "46749e91-fd6f-4dcd-8c10-c11352de9138" (UID: "46749e91-fd6f-4dcd-8c10-c11352de9138"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500085 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run" (OuterVolumeSpecName: "var-run") pod "46749e91-fd6f-4dcd-8c10-c11352de9138" (UID: "46749e91-fd6f-4dcd-8c10-c11352de9138"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500131 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-additional-scripts\") pod \"46749e91-fd6f-4dcd-8c10-c11352de9138\" (UID: \"46749e91-fd6f-4dcd-8c10-c11352de9138\") " Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500518 4710 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500533 4710 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500542 4710 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/46749e91-fd6f-4dcd-8c10-c11352de9138-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500657 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "46749e91-fd6f-4dcd-8c10-c11352de9138" (UID: "46749e91-fd6f-4dcd-8c10-c11352de9138"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.500915 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-scripts" (OuterVolumeSpecName: "scripts") pod "46749e91-fd6f-4dcd-8c10-c11352de9138" (UID: "46749e91-fd6f-4dcd-8c10-c11352de9138"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.503848 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46749e91-fd6f-4dcd-8c10-c11352de9138-kube-api-access-ww9ms" (OuterVolumeSpecName: "kube-api-access-ww9ms") pod "46749e91-fd6f-4dcd-8c10-c11352de9138" (UID: "46749e91-fd6f-4dcd-8c10-c11352de9138"). InnerVolumeSpecName "kube-api-access-ww9ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.514823 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.602924 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww9ms\" (UniqueName: \"kubernetes.io/projected/46749e91-fd6f-4dcd-8c10-c11352de9138-kube-api-access-ww9ms\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.602956 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.602968 4710 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/46749e91-fd6f-4dcd-8c10-c11352de9138-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.964819 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fp7wl-config-wq2tg" event={"ID":"46749e91-fd6f-4dcd-8c10-c11352de9138","Type":"ContainerDied","Data":"979762defae71800b9a87edabc5118677239dd47e25fc1e7e19f6bbb133f37bf"} Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.965103 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="979762defae71800b9a87edabc5118677239dd47e25fc1e7e19f6bbb133f37bf" Oct 02 13:06:33 crc kubenswrapper[4710]: I1002 13:06:33.965178 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fp7wl-config-wq2tg" Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.285328 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.412757 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fp7wl-config-wq2tg"] Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.419261 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fp7wl-config-wq2tg"] Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.473121 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.683861 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.684180 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="prometheus" containerID="cri-o://17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb" gracePeriod=600 Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.684260 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="thanos-sidecar" containerID="cri-o://896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768" gracePeriod=600 Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.684348 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="config-reloader" containerID="cri-o://7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6" gracePeriod=600 Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.736607 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="98c2f825-d5c6-47e7-a96d-8cf7227b0a24" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.916053 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46749e91-fd6f-4dcd-8c10-c11352de9138" path="/var/lib/kubelet/pods/46749e91-fd6f-4dcd-8c10-c11352de9138/volumes" Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.983393 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"a59cb1af6d4b27e573673f7e5c4aa5360769e332493046d37e55742e74269bfb"} Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.988351 4710 generic.go:334] "Generic (PLEG): container finished" podID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerID="896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768" exitCode=0 Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.988377 4710 generic.go:334] "Generic (PLEG): container finished" podID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerID="17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb" exitCode=0 Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.988586 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerDied","Data":"896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768"} Oct 02 13:06:34 crc kubenswrapper[4710]: I1002 13:06:34.988614 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerDied","Data":"17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb"} Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.106774 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.544356 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.642977 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-274w2\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-kube-api-access-274w2\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643134 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643158 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8530440e-9b9d-40bc-95ac-b7c40899b0cb-prometheus-metric-storage-rulefiles-0\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643190 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-web-config\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643236 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643583 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config-out\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643623 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-tls-assets\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.643648 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-thanos-prometheus-http-client-file\") pod \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\" (UID: \"8530440e-9b9d-40bc-95ac-b7c40899b0cb\") " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.644824 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8530440e-9b9d-40bc-95ac-b7c40899b0cb-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.649511 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-kube-api-access-274w2" (OuterVolumeSpecName: "kube-api-access-274w2") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "kube-api-access-274w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.649921 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config" (OuterVolumeSpecName: "config") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.650392 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.650826 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config-out" (OuterVolumeSpecName: "config-out") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.650946 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.673849 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "pvc-11c95985-a270-4e8f-9943-0f6910c78dc6". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.683959 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-web-config" (OuterVolumeSpecName: "web-config") pod "8530440e-9b9d-40bc-95ac-b7c40899b0cb" (UID: "8530440e-9b9d-40bc-95ac-b7c40899b0cb"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746434 4710 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746478 4710 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746491 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-274w2\" (UniqueName: \"kubernetes.io/projected/8530440e-9b9d-40bc-95ac-b7c40899b0cb-kube-api-access-274w2\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746533 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") on node \"crc\" " Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746545 4710 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8530440e-9b9d-40bc-95ac-b7c40899b0cb-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746555 4710 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-web-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746564 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.746572 4710 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8530440e-9b9d-40bc-95ac-b7c40899b0cb-config-out\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.766324 4710 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.766476 4710 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-11c95985-a270-4e8f-9943-0f6910c78dc6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6") on node "crc" Oct 02 13:06:35 crc kubenswrapper[4710]: I1002 13:06:35.848884 4710 reconciler_common.go:293] "Volume detached for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.006432 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"bda2945cad53f1d8eafd2f5cf016e80788385fd089a2b88775cc0426c424f547"} Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.010295 4710 generic.go:334] "Generic (PLEG): container finished" podID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerID="7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6" exitCode=0 Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.010338 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerDied","Data":"7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6"} Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.010365 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.010375 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8530440e-9b9d-40bc-95ac-b7c40899b0cb","Type":"ContainerDied","Data":"547830786af86977452d67e2d13a17b0a09dda336890f14773836cf91ac87bb6"} Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.010421 4710 scope.go:117] "RemoveContainer" containerID="896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.044527 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.045299 4710 scope.go:117] "RemoveContainer" containerID="7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.050936 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076304 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.076738 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="prometheus" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076776 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="prometheus" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.076801 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="config-reloader" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076808 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="config-reloader" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.076830 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="init-config-reloader" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076837 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="init-config-reloader" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.076853 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46749e91-fd6f-4dcd-8c10-c11352de9138" containerName="ovn-config" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076860 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="46749e91-fd6f-4dcd-8c10-c11352de9138" containerName="ovn-config" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.076870 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce991bb5-1e98-40f5-8395-b1997433b236" containerName="mariadb-account-create" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076876 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce991bb5-1e98-40f5-8395-b1997433b236" containerName="mariadb-account-create" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.076889 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="thanos-sidecar" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.076895 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="thanos-sidecar" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.077042 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="46749e91-fd6f-4dcd-8c10-c11352de9138" containerName="ovn-config" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.077054 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce991bb5-1e98-40f5-8395-b1997433b236" containerName="mariadb-account-create" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.077065 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="prometheus" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.077077 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="config-reloader" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.077085 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" containerName="thanos-sidecar" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.078778 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.085821 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.086162 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.086626 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.087482 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.087705 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-58df2" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.087927 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.095317 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.100781 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.105941 4710 scope.go:117] "RemoveContainer" containerID="17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.140725 4710 scope.go:117] "RemoveContainer" containerID="3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.187942 4710 scope.go:117] "RemoveContainer" containerID="896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.188678 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768\": container with ID starting with 896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768 not found: ID does not exist" containerID="896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.188717 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768"} err="failed to get container status \"896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768\": rpc error: code = NotFound desc = could not find container \"896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768\": container with ID starting with 896fd7dc10d5126ddd78189dfe5410ee2a2b7ca41c633c3939be61eafa90e768 not found: ID does not exist" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.188821 4710 scope.go:117] "RemoveContainer" containerID="7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.189893 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6\": container with ID starting with 7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6 not found: ID does not exist" containerID="7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.189941 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6"} err="failed to get container status \"7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6\": rpc error: code = NotFound desc = could not find container \"7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6\": container with ID starting with 7252854e6ffeac6f1c2d53fb877661542ca0d75bad5f03e2d14d64b895850cb6 not found: ID does not exist" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.189967 4710 scope.go:117] "RemoveContainer" containerID="17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.194886 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb\": container with ID starting with 17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb not found: ID does not exist" containerID="17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.194941 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb"} err="failed to get container status \"17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb\": rpc error: code = NotFound desc = could not find container \"17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb\": container with ID starting with 17a97771eab87b0f69c5a2395631ee5fa5a139b219a7d9a62be9239e8d3b26eb not found: ID does not exist" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.194974 4710 scope.go:117] "RemoveContainer" containerID="3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e" Oct 02 13:06:36 crc kubenswrapper[4710]: E1002 13:06:36.195721 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e\": container with ID starting with 3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e not found: ID does not exist" containerID="3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.195771 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e"} err="failed to get container status \"3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e\": rpc error: code = NotFound desc = could not find container \"3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e\": container with ID starting with 3331e83cead349fa61986833c8e89e648b7ee4193f406c3488fbfe0457552d8e not found: ID does not exist" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.257617 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.257676 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.257844 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.257931 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258008 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmkcj\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-kube-api-access-mmkcj\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258058 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258189 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258335 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258370 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258402 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.258558 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360299 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360385 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360421 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360465 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360496 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmkcj\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-kube-api-access-mmkcj\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360517 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360560 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360599 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360618 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360638 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.360676 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.363982 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.366713 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.367455 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.367485 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.367673 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.372802 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.372920 4710 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.372925 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.372955 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/549ec1f26b5b9c00df36a940184c18d591918f74b3d89fea9d49d0eb0309560f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.373184 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.373386 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.383688 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmkcj\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-kube-api-access-mmkcj\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.430813 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:36 crc kubenswrapper[4710]: I1002 13:06:36.702091 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:37 crc kubenswrapper[4710]: I1002 13:06:37.034320 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8530440e-9b9d-40bc-95ac-b7c40899b0cb" path="/var/lib/kubelet/pods/8530440e-9b9d-40bc-95ac-b7c40899b0cb/volumes" Oct 02 13:06:37 crc kubenswrapper[4710]: I1002 13:06:37.035786 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"7e7bb14704463527c15a0325a2631ec290f31dfec56a687e84d7125c106c5549"} Oct 02 13:06:37 crc kubenswrapper[4710]: I1002 13:06:37.035819 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"58140649c199c4d341453cd7b0e4d55aa7a85e19f7ea73b81a391653a0fcbc54"} Oct 02 13:06:37 crc kubenswrapper[4710]: I1002 13:06:37.035838 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"e7d2df16066153b31d6f5fabea8817643cb1111fad6b1fb09670dc36b774c982"} Oct 02 13:06:37 crc kubenswrapper[4710]: I1002 13:06:37.194702 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:06:38 crc kubenswrapper[4710]: I1002 13:06:38.033105 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerStarted","Data":"ecf2c4b319af9b9600dc61fbf2cadcd99efdb00314a72310f645c7e8374462a3"} Oct 02 13:06:41 crc kubenswrapper[4710]: I1002 13:06:41.058263 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerStarted","Data":"9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6"} Oct 02 13:06:44 crc kubenswrapper[4710]: I1002 13:06:44.472996 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:06:44 crc kubenswrapper[4710]: I1002 13:06:44.737012 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Oct 02 13:06:45 crc kubenswrapper[4710]: I1002 13:06:45.105002 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 13:06:45 crc kubenswrapper[4710]: I1002 13:06:45.122457 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"173a606426fc043c93f200985691f2d83424e07847fa590d78b4591f88c1c346"} Oct 02 13:06:46 crc kubenswrapper[4710]: I1002 13:06:46.155448 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6htbd" event={"ID":"52ce5f44-dff4-4fe5-aa15-e76e352b9d88","Type":"ContainerStarted","Data":"78a723072bd56a19fb6d40a5635d3457b5b60090db8f1c0692d0cd2d73db4adf"} Oct 02 13:06:46 crc kubenswrapper[4710]: I1002 13:06:46.159112 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"7cd53ad6ffe13cbc7389fbd705518a5c7945d07b977ad2c9343a142d6c133714"} Oct 02 13:06:46 crc kubenswrapper[4710]: I1002 13:06:46.159162 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"ee2cde82664e1a9fc8c1c94cff184f04ea18f77684c86fce674bd8f6769accdf"} Oct 02 13:06:46 crc kubenswrapper[4710]: I1002 13:06:46.159173 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"31277103f3c80b7341dd738c69fc3e685f690d614ca95dae93fb4e8ec303146d"} Oct 02 13:06:46 crc kubenswrapper[4710]: I1002 13:06:46.176052 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-6htbd" podStartSLOduration=3.166268468 podStartE2EDuration="18.17603266s" podCreationTimestamp="2025-10-02 13:06:28 +0000 UTC" firstStartedPulling="2025-10-02 13:06:29.814033472 +0000 UTC m=+1033.920444355" lastFinishedPulling="2025-10-02 13:06:44.823797654 +0000 UTC m=+1048.930208547" observedRunningTime="2025-10-02 13:06:46.173338072 +0000 UTC m=+1050.279748955" watchObservedRunningTime="2025-10-02 13:06:46.17603266 +0000 UTC m=+1050.282443543" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.195482 4710 generic.go:334] "Generic (PLEG): container finished" podID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerID="9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6" exitCode=0 Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.195547 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerDied","Data":"9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6"} Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.228343 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"0675cdb3fc5c526a370af57b367ea575ee01447e29b6eb4429c1b405acd24f59"} Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.228382 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"7ed1734a98fb46325539b046bb5eaf966dc130c77666a28f60cd6351f7d5b87e"} Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.228392 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"1d9c4a5c9d58e05810440a57cdbb34f3a76e2b52831c119e8f6adaf224da6e52"} Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.228401 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"5c7121611e0d608de408ed84e6857f29f30fd9420212156a5d202f694c4a861b"} Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.421545 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-cw6sh"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.423057 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.448232 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cw6sh"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.449637 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxtfv\" (UniqueName: \"kubernetes.io/projected/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe-kube-api-access-hxtfv\") pod \"cinder-db-create-cw6sh\" (UID: \"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe\") " pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.522916 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-b4nsw"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.524461 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.556775 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b4nsw"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.559835 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbg7f\" (UniqueName: \"kubernetes.io/projected/25de8c63-18f4-4896-b856-3c5c20d2d54a-kube-api-access-mbg7f\") pod \"barbican-db-create-b4nsw\" (UID: \"25de8c63-18f4-4896-b856-3c5c20d2d54a\") " pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.559928 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxtfv\" (UniqueName: \"kubernetes.io/projected/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe-kube-api-access-hxtfv\") pod \"cinder-db-create-cw6sh\" (UID: \"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe\") " pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.627458 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxtfv\" (UniqueName: \"kubernetes.io/projected/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe-kube-api-access-hxtfv\") pod \"cinder-db-create-cw6sh\" (UID: \"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe\") " pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.670710 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbg7f\" (UniqueName: \"kubernetes.io/projected/25de8c63-18f4-4896-b856-3c5c20d2d54a-kube-api-access-mbg7f\") pod \"barbican-db-create-b4nsw\" (UID: \"25de8c63-18f4-4896-b856-3c5c20d2d54a\") " pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.713379 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbg7f\" (UniqueName: \"kubernetes.io/projected/25de8c63-18f4-4896-b856-3c5c20d2d54a-kube-api-access-mbg7f\") pod \"barbican-db-create-b4nsw\" (UID: \"25de8c63-18f4-4896-b856-3c5c20d2d54a\") " pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.713436 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-cm5dq"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.727253 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.738349 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-pvpl2" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.738865 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-cm5dq"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.738901 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.768584 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.821318 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xf9wb"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.822619 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.826082 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.826273 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.826405 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.826578 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p52kv" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.852119 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xf9wb"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.874039 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-db-sync-config-data\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.874099 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-config-data\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.874118 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-combined-ca-bundle\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.874152 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpbs7\" (UniqueName: \"kubernetes.io/projected/765efee7-b0c7-417f-8ad6-fee3d089bc50-kube-api-access-jpbs7\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.892826 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-wc99n"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.894208 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.902174 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wc99n"] Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.911975 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978164 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpbs7\" (UniqueName: \"kubernetes.io/projected/765efee7-b0c7-417f-8ad6-fee3d089bc50-kube-api-access-jpbs7\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978464 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jftj\" (UniqueName: \"kubernetes.io/projected/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-kube-api-access-7jftj\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978554 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-config-data\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978715 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-combined-ca-bundle\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978808 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-db-sync-config-data\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978910 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-config-data\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.978992 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-combined-ca-bundle\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.979086 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsgbx\" (UniqueName: \"kubernetes.io/projected/79a4c671-5a25-4cd5-9b19-db54e931b28b-kube-api-access-fsgbx\") pod \"neutron-db-create-wc99n\" (UID: \"79a4c671-5a25-4cd5-9b19-db54e931b28b\") " pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.986049 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-config-data\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.998239 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-db-sync-config-data\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:47 crc kubenswrapper[4710]: I1002 13:06:47.999522 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpbs7\" (UniqueName: \"kubernetes.io/projected/765efee7-b0c7-417f-8ad6-fee3d089bc50-kube-api-access-jpbs7\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.001368 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-combined-ca-bundle\") pod \"watcher-db-sync-cm5dq\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.081242 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jftj\" (UniqueName: \"kubernetes.io/projected/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-kube-api-access-7jftj\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.081291 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-config-data\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.081371 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-combined-ca-bundle\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.081425 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsgbx\" (UniqueName: \"kubernetes.io/projected/79a4c671-5a25-4cd5-9b19-db54e931b28b-kube-api-access-fsgbx\") pod \"neutron-db-create-wc99n\" (UID: \"79a4c671-5a25-4cd5-9b19-db54e931b28b\") " pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.089935 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.096245 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-config-data\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.100421 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-combined-ca-bundle\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.125487 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jftj\" (UniqueName: \"kubernetes.io/projected/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-kube-api-access-7jftj\") pod \"keystone-db-sync-xf9wb\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.131973 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsgbx\" (UniqueName: \"kubernetes.io/projected/79a4c671-5a25-4cd5-9b19-db54e931b28b-kube-api-access-fsgbx\") pod \"neutron-db-create-wc99n\" (UID: \"79a4c671-5a25-4cd5-9b19-db54e931b28b\") " pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.173635 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.224674 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.335575 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"66a9fd039846a080b568e41456f4e42191a740d8b24f0628914bc1e2ba7fc990"} Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.335621 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"d3d859ddd145195dde15173c8ea3674b2cdca4d038ea8276cfd557e1b9b7da84"} Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.353302 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerStarted","Data":"c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88"} Oct 02 13:06:48 crc kubenswrapper[4710]: W1002 13:06:48.480376 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fa748a4_55a5_4f2b_99f9_6d049d7c4bbe.slice/crio-d8868d28ab8f182313006d183b8faaa162303df3f486a1016754fde45f200349 WatchSource:0}: Error finding container d8868d28ab8f182313006d183b8faaa162303df3f486a1016754fde45f200349: Status 404 returned error can't find the container with id d8868d28ab8f182313006d183b8faaa162303df3f486a1016754fde45f200349 Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.481079 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cw6sh"] Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.617434 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b4nsw"] Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.746872 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-cm5dq"] Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.894990 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xf9wb"] Oct 02 13:06:48 crc kubenswrapper[4710]: W1002 13:06:48.921866 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad3b2f5f_f4d4_4a80_948d_275033ab9f05.slice/crio-287ab3ae101f86b43a0cca33cc6f5f4c7e02f3e87443b18ff38841208f1b672a WatchSource:0}: Error finding container 287ab3ae101f86b43a0cca33cc6f5f4c7e02f3e87443b18ff38841208f1b672a: Status 404 returned error can't find the container with id 287ab3ae101f86b43a0cca33cc6f5f4c7e02f3e87443b18ff38841208f1b672a Oct 02 13:06:48 crc kubenswrapper[4710]: I1002 13:06:48.994030 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wc99n"] Oct 02 13:06:49 crc kubenswrapper[4710]: W1002 13:06:49.019774 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79a4c671_5a25_4cd5_9b19_db54e931b28b.slice/crio-d24db208180f4cec12a9b30a3020525a018f91a5dd2f960d5a65c1072aac3a19 WatchSource:0}: Error finding container d24db208180f4cec12a9b30a3020525a018f91a5dd2f960d5a65c1072aac3a19: Status 404 returned error can't find the container with id d24db208180f4cec12a9b30a3020525a018f91a5dd2f960d5a65c1072aac3a19 Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.366134 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf9wb" event={"ID":"ad3b2f5f-f4d4-4a80-948d-275033ab9f05","Type":"ContainerStarted","Data":"287ab3ae101f86b43a0cca33cc6f5f4c7e02f3e87443b18ff38841208f1b672a"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.366477 4710 generic.go:334] "Generic (PLEG): container finished" podID="25de8c63-18f4-4896-b856-3c5c20d2d54a" containerID="2b30bf42c58fda06f2f6f5367a2b184adbcbbe45eb40d08292630d7659fef79e" exitCode=0 Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.366543 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4nsw" event={"ID":"25de8c63-18f4-4896-b856-3c5c20d2d54a","Type":"ContainerDied","Data":"2b30bf42c58fda06f2f6f5367a2b184adbcbbe45eb40d08292630d7659fef79e"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.366569 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4nsw" event={"ID":"25de8c63-18f4-4896-b856-3c5c20d2d54a","Type":"ContainerStarted","Data":"43634ceecfc03844e7e5778da8728bf76eb039e2569c108d5ab2afb39703525a"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.373326 4710 generic.go:334] "Generic (PLEG): container finished" podID="2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe" containerID="4a41bdc2ba6285c7733c4b96675ea15361dc0b06aaf8cb2059defb06e978fa4d" exitCode=0 Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.373477 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cw6sh" event={"ID":"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe","Type":"ContainerDied","Data":"4a41bdc2ba6285c7733c4b96675ea15361dc0b06aaf8cb2059defb06e978fa4d"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.373501 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cw6sh" event={"ID":"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe","Type":"ContainerStarted","Data":"d8868d28ab8f182313006d183b8faaa162303df3f486a1016754fde45f200349"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.415207 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"67031d3a-15f2-4b86-92d7-b525a3a8420e","Type":"ContainerStarted","Data":"bdfab64e8e19ede2a189968f75ae850989e3a0b7e519b75c307f97e6cae5fa79"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.419003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wc99n" event={"ID":"79a4c671-5a25-4cd5-9b19-db54e931b28b","Type":"ContainerStarted","Data":"2a3b4974ea13ea15af50dfac12ed9bf8640b4b508b0ab89e17d5f03c364b5a17"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.419040 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wc99n" event={"ID":"79a4c671-5a25-4cd5-9b19-db54e931b28b","Type":"ContainerStarted","Data":"d24db208180f4cec12a9b30a3020525a018f91a5dd2f960d5a65c1072aac3a19"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.420198 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-cm5dq" event={"ID":"765efee7-b0c7-417f-8ad6-fee3d089bc50","Type":"ContainerStarted","Data":"00ce7925a7fc416b0244d667b9e958862aa127280db00db3278ec17d0c20df60"} Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.466107 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.382045193 podStartE2EDuration="49.460960042s" podCreationTimestamp="2025-10-02 13:06:00 +0000 UTC" firstStartedPulling="2025-10-02 13:06:34.310545423 +0000 UTC m=+1038.416956306" lastFinishedPulling="2025-10-02 13:06:46.389460272 +0000 UTC m=+1050.495871155" observedRunningTime="2025-10-02 13:06:49.450636362 +0000 UTC m=+1053.557047255" watchObservedRunningTime="2025-10-02 13:06:49.460960042 +0000 UTC m=+1053.567370925" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.697798 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d9c777fc7-mjjbd"] Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.699312 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.704035 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.707229 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d9c777fc7-mjjbd"] Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.819659 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-svc\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.819716 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-nb\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.819773 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66n2t\" (UniqueName: \"kubernetes.io/projected/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-kube-api-access-66n2t\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.819855 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-config\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.819893 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-swift-storage-0\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.819928 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-sb\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.922658 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-sb\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.922764 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-svc\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.922786 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-nb\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.922819 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66n2t\" (UniqueName: \"kubernetes.io/projected/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-kube-api-access-66n2t\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.922885 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-config\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.922912 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-swift-storage-0\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.923867 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-swift-storage-0\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.923904 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-nb\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.923905 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-svc\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.923952 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-config\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.924562 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-sb\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:49 crc kubenswrapper[4710]: I1002 13:06:49.950941 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66n2t\" (UniqueName: \"kubernetes.io/projected/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-kube-api-access-66n2t\") pod \"dnsmasq-dns-d9c777fc7-mjjbd\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:50 crc kubenswrapper[4710]: I1002 13:06:50.036342 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:06:50 crc kubenswrapper[4710]: I1002 13:06:50.494279 4710 generic.go:334] "Generic (PLEG): container finished" podID="79a4c671-5a25-4cd5-9b19-db54e931b28b" containerID="2a3b4974ea13ea15af50dfac12ed9bf8640b4b508b0ab89e17d5f03c364b5a17" exitCode=0 Oct 02 13:06:50 crc kubenswrapper[4710]: I1002 13:06:50.494589 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wc99n" event={"ID":"79a4c671-5a25-4cd5-9b19-db54e931b28b","Type":"ContainerDied","Data":"2a3b4974ea13ea15af50dfac12ed9bf8640b4b508b0ab89e17d5f03c364b5a17"} Oct 02 13:06:50 crc kubenswrapper[4710]: I1002 13:06:50.519690 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerStarted","Data":"334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0"} Oct 02 13:06:50 crc kubenswrapper[4710]: I1002 13:06:50.583110 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d9c777fc7-mjjbd"] Oct 02 13:06:50 crc kubenswrapper[4710]: I1002 13:06:50.899044 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.053040 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.056286 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbg7f\" (UniqueName: \"kubernetes.io/projected/25de8c63-18f4-4896-b856-3c5c20d2d54a-kube-api-access-mbg7f\") pod \"25de8c63-18f4-4896-b856-3c5c20d2d54a\" (UID: \"25de8c63-18f4-4896-b856-3c5c20d2d54a\") " Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.062006 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25de8c63-18f4-4896-b856-3c5c20d2d54a-kube-api-access-mbg7f" (OuterVolumeSpecName: "kube-api-access-mbg7f") pod "25de8c63-18f4-4896-b856-3c5c20d2d54a" (UID: "25de8c63-18f4-4896-b856-3c5c20d2d54a"). InnerVolumeSpecName "kube-api-access-mbg7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.062952 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.166945 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxtfv\" (UniqueName: \"kubernetes.io/projected/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe-kube-api-access-hxtfv\") pod \"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe\" (UID: \"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe\") " Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.167079 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsgbx\" (UniqueName: \"kubernetes.io/projected/79a4c671-5a25-4cd5-9b19-db54e931b28b-kube-api-access-fsgbx\") pod \"79a4c671-5a25-4cd5-9b19-db54e931b28b\" (UID: \"79a4c671-5a25-4cd5-9b19-db54e931b28b\") " Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.167480 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbg7f\" (UniqueName: \"kubernetes.io/projected/25de8c63-18f4-4896-b856-3c5c20d2d54a-kube-api-access-mbg7f\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.171488 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe-kube-api-access-hxtfv" (OuterVolumeSpecName: "kube-api-access-hxtfv") pod "2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe" (UID: "2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe"). InnerVolumeSpecName "kube-api-access-hxtfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.183199 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a4c671-5a25-4cd5-9b19-db54e931b28b-kube-api-access-fsgbx" (OuterVolumeSpecName: "kube-api-access-fsgbx") pod "79a4c671-5a25-4cd5-9b19-db54e931b28b" (UID: "79a4c671-5a25-4cd5-9b19-db54e931b28b"). InnerVolumeSpecName "kube-api-access-fsgbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.269354 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxtfv\" (UniqueName: \"kubernetes.io/projected/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe-kube-api-access-hxtfv\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.269398 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsgbx\" (UniqueName: \"kubernetes.io/projected/79a4c671-5a25-4cd5-9b19-db54e931b28b-kube-api-access-fsgbx\") on node \"crc\" DevicePath \"\"" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.530832 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4nsw" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.530514 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4nsw" event={"ID":"25de8c63-18f4-4896-b856-3c5c20d2d54a","Type":"ContainerDied","Data":"43634ceecfc03844e7e5778da8728bf76eb039e2569c108d5ab2afb39703525a"} Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.532308 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43634ceecfc03844e7e5778da8728bf76eb039e2569c108d5ab2afb39703525a" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.535187 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cw6sh" event={"ID":"2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe","Type":"ContainerDied","Data":"d8868d28ab8f182313006d183b8faaa162303df3f486a1016754fde45f200349"} Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.535249 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8868d28ab8f182313006d183b8faaa162303df3f486a1016754fde45f200349" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.535299 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cw6sh" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.541851 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerStarted","Data":"5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271"} Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.544646 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wc99n" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.544645 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wc99n" event={"ID":"79a4c671-5a25-4cd5-9b19-db54e931b28b","Type":"ContainerDied","Data":"d24db208180f4cec12a9b30a3020525a018f91a5dd2f960d5a65c1072aac3a19"} Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.545840 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d24db208180f4cec12a9b30a3020525a018f91a5dd2f960d5a65c1072aac3a19" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.549564 4710 generic.go:334] "Generic (PLEG): container finished" podID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerID="8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a" exitCode=0 Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.549617 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" event={"ID":"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507","Type":"ContainerDied","Data":"8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a"} Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.549635 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" event={"ID":"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507","Type":"ContainerStarted","Data":"63e27bf82ff788017f8ea9ece832018a26b0af1d8e6c93db08a8d5ed81f2bc61"} Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.573195 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=15.573177257 podStartE2EDuration="15.573177257s" podCreationTimestamp="2025-10-02 13:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:06:51.57250468 +0000 UTC m=+1055.678915583" watchObservedRunningTime="2025-10-02 13:06:51.573177257 +0000 UTC m=+1055.679588140" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.702707 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.702816 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:51 crc kubenswrapper[4710]: I1002 13:06:51.709284 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:52 crc kubenswrapper[4710]: I1002 13:06:52.563693 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.549047 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-a0e6-account-create-n8wcs"] Oct 02 13:06:57 crc kubenswrapper[4710]: E1002 13:06:57.550199 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.550214 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: E1002 13:06:57.550248 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25de8c63-18f4-4896-b856-3c5c20d2d54a" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.550256 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="25de8c63-18f4-4896-b856-3c5c20d2d54a" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: E1002 13:06:57.550269 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a4c671-5a25-4cd5-9b19-db54e931b28b" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.550278 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a4c671-5a25-4cd5-9b19-db54e931b28b" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.550445 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.550462 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a4c671-5a25-4cd5-9b19-db54e931b28b" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.550478 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="25de8c63-18f4-4896-b856-3c5c20d2d54a" containerName="mariadb-database-create" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.551201 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.553210 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.557024 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a0e6-account-create-n8wcs"] Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.647303 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-21d4-account-create-r2d77"] Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.648583 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.653512 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.667233 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-21d4-account-create-r2d77"] Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.691548 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tt54\" (UniqueName: \"kubernetes.io/projected/34948622-b486-47f0-bbb7-832f5ca6a1e7-kube-api-access-5tt54\") pod \"cinder-a0e6-account-create-n8wcs\" (UID: \"34948622-b486-47f0-bbb7-832f5ca6a1e7\") " pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.793295 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tt54\" (UniqueName: \"kubernetes.io/projected/34948622-b486-47f0-bbb7-832f5ca6a1e7-kube-api-access-5tt54\") pod \"cinder-a0e6-account-create-n8wcs\" (UID: \"34948622-b486-47f0-bbb7-832f5ca6a1e7\") " pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.793391 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svf5f\" (UniqueName: \"kubernetes.io/projected/39dc6a17-1fa2-440c-9ee4-3b12755fc422-kube-api-access-svf5f\") pod \"barbican-21d4-account-create-r2d77\" (UID: \"39dc6a17-1fa2-440c-9ee4-3b12755fc422\") " pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.815019 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tt54\" (UniqueName: \"kubernetes.io/projected/34948622-b486-47f0-bbb7-832f5ca6a1e7-kube-api-access-5tt54\") pod \"cinder-a0e6-account-create-n8wcs\" (UID: \"34948622-b486-47f0-bbb7-832f5ca6a1e7\") " pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.869986 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9aae-account-create-vt2hh"] Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.871263 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.873704 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.874776 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.878868 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9aae-account-create-vt2hh"] Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.895133 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svf5f\" (UniqueName: \"kubernetes.io/projected/39dc6a17-1fa2-440c-9ee4-3b12755fc422-kube-api-access-svf5f\") pod \"barbican-21d4-account-create-r2d77\" (UID: \"39dc6a17-1fa2-440c-9ee4-3b12755fc422\") " pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.911958 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svf5f\" (UniqueName: \"kubernetes.io/projected/39dc6a17-1fa2-440c-9ee4-3b12755fc422-kube-api-access-svf5f\") pod \"barbican-21d4-account-create-r2d77\" (UID: \"39dc6a17-1fa2-440c-9ee4-3b12755fc422\") " pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.975168 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:06:57 crc kubenswrapper[4710]: I1002 13:06:57.998214 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcnsv\" (UniqueName: \"kubernetes.io/projected/03930039-3e8f-4a37-ae83-61f08a70aea4-kube-api-access-dcnsv\") pod \"neutron-9aae-account-create-vt2hh\" (UID: \"03930039-3e8f-4a37-ae83-61f08a70aea4\") " pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:06:58 crc kubenswrapper[4710]: I1002 13:06:58.099769 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcnsv\" (UniqueName: \"kubernetes.io/projected/03930039-3e8f-4a37-ae83-61f08a70aea4-kube-api-access-dcnsv\") pod \"neutron-9aae-account-create-vt2hh\" (UID: \"03930039-3e8f-4a37-ae83-61f08a70aea4\") " pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:06:58 crc kubenswrapper[4710]: I1002 13:06:58.120449 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcnsv\" (UniqueName: \"kubernetes.io/projected/03930039-3e8f-4a37-ae83-61f08a70aea4-kube-api-access-dcnsv\") pod \"neutron-9aae-account-create-vt2hh\" (UID: \"03930039-3e8f-4a37-ae83-61f08a70aea4\") " pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:06:58 crc kubenswrapper[4710]: I1002 13:06:58.200047 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:06:59 crc kubenswrapper[4710]: I1002 13:06:59.625209 4710 generic.go:334] "Generic (PLEG): container finished" podID="52ce5f44-dff4-4fe5-aa15-e76e352b9d88" containerID="78a723072bd56a19fb6d40a5635d3457b5b60090db8f1c0692d0cd2d73db4adf" exitCode=0 Oct 02 13:06:59 crc kubenswrapper[4710]: I1002 13:06:59.625286 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6htbd" event={"ID":"52ce5f44-dff4-4fe5-aa15-e76e352b9d88","Type":"ContainerDied","Data":"78a723072bd56a19fb6d40a5635d3457b5b60090db8f1c0692d0cd2d73db4adf"} Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.534656 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6htbd" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.662556 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf9wb" event={"ID":"ad3b2f5f-f4d4-4a80-948d-275033ab9f05","Type":"ContainerStarted","Data":"23d65eded6f15a0f22f1963e12678f7c593c49010d2f67512979d07691a537a6"} Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.670566 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6htbd" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.670619 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6htbd" event={"ID":"52ce5f44-dff4-4fe5-aa15-e76e352b9d88","Type":"ContainerDied","Data":"94ff6ca2311527d537d0ef953dcad1588e42d109c3e7e0b702884907cc5d0a75"} Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.670670 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94ff6ca2311527d537d0ef953dcad1588e42d109c3e7e0b702884907cc5d0a75" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.676129 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.679789 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-combined-ca-bundle\") pod \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.680333 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2w26\" (UniqueName: \"kubernetes.io/projected/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-kube-api-access-m2w26\") pod \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.680401 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-db-sync-config-data\") pod \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.680446 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-config-data\") pod \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\" (UID: \"52ce5f44-dff4-4fe5-aa15-e76e352b9d88\") " Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.688942 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-kube-api-access-m2w26" (OuterVolumeSpecName: "kube-api-access-m2w26") pod "52ce5f44-dff4-4fe5-aa15-e76e352b9d88" (UID: "52ce5f44-dff4-4fe5-aa15-e76e352b9d88"). InnerVolumeSpecName "kube-api-access-m2w26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.690031 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "52ce5f44-dff4-4fe5-aa15-e76e352b9d88" (UID: "52ce5f44-dff4-4fe5-aa15-e76e352b9d88"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.703438 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-cm5dq" podStartSLOduration=2.150964418 podStartE2EDuration="14.703417427s" podCreationTimestamp="2025-10-02 13:06:47 +0000 UTC" firstStartedPulling="2025-10-02 13:06:48.756443299 +0000 UTC m=+1052.862854182" lastFinishedPulling="2025-10-02 13:07:01.308896308 +0000 UTC m=+1065.415307191" observedRunningTime="2025-10-02 13:07:01.702313539 +0000 UTC m=+1065.808724422" watchObservedRunningTime="2025-10-02 13:07:01.703417427 +0000 UTC m=+1065.809828330" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.705619 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xf9wb" podStartSLOduration=2.378328061 podStartE2EDuration="14.705608772s" podCreationTimestamp="2025-10-02 13:06:47 +0000 UTC" firstStartedPulling="2025-10-02 13:06:48.934107901 +0000 UTC m=+1053.040518784" lastFinishedPulling="2025-10-02 13:07:01.261388612 +0000 UTC m=+1065.367799495" observedRunningTime="2025-10-02 13:07:01.685493206 +0000 UTC m=+1065.791904089" watchObservedRunningTime="2025-10-02 13:07:01.705608772 +0000 UTC m=+1065.812019655" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.717605 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52ce5f44-dff4-4fe5-aa15-e76e352b9d88" (UID: "52ce5f44-dff4-4fe5-aa15-e76e352b9d88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.740190 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a0e6-account-create-n8wcs"] Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.743189 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" podStartSLOduration=12.743167028 podStartE2EDuration="12.743167028s" podCreationTimestamp="2025-10-02 13:06:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:01.718959378 +0000 UTC m=+1065.825370261" watchObservedRunningTime="2025-10-02 13:07:01.743167028 +0000 UTC m=+1065.849577911" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.761964 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-config-data" (OuterVolumeSpecName: "config-data") pod "52ce5f44-dff4-4fe5-aa15-e76e352b9d88" (UID: "52ce5f44-dff4-4fe5-aa15-e76e352b9d88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.782849 4710 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.782887 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.782897 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.782906 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2w26\" (UniqueName: \"kubernetes.io/projected/52ce5f44-dff4-4fe5-aa15-e76e352b9d88-kube-api-access-m2w26\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.827897 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9aae-account-create-vt2hh"] Oct 02 13:07:01 crc kubenswrapper[4710]: I1002 13:07:01.956698 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-21d4-account-create-r2d77"] Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.042840 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d9c777fc7-mjjbd"] Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.096667 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89fc65cf9-zg284"] Oct 02 13:07:02 crc kubenswrapper[4710]: E1002 13:07:02.097117 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52ce5f44-dff4-4fe5-aa15-e76e352b9d88" containerName="glance-db-sync" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.097129 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="52ce5f44-dff4-4fe5-aa15-e76e352b9d88" containerName="glance-db-sync" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.097295 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="52ce5f44-dff4-4fe5-aa15-e76e352b9d88" containerName="glance-db-sync" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.098360 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.109001 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89fc65cf9-zg284"] Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.207001 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-swift-storage-0\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.207057 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wdld\" (UniqueName: \"kubernetes.io/projected/64d612e0-1f54-4609-a642-bd78fd3a1269-kube-api-access-9wdld\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.207236 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-config\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.207323 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-nb\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.207368 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-svc\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.207448 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-sb\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.309177 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-config\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.310539 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-config\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.310621 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-nb\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.310714 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-svc\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.311568 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-nb\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.311644 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-svc\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.311877 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-sb\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.312559 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-sb\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.312929 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-swift-storage-0\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.312991 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wdld\" (UniqueName: \"kubernetes.io/projected/64d612e0-1f54-4609-a642-bd78fd3a1269-kube-api-access-9wdld\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.313892 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-swift-storage-0\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.344615 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wdld\" (UniqueName: \"kubernetes.io/projected/64d612e0-1f54-4609-a642-bd78fd3a1269-kube-api-access-9wdld\") pod \"dnsmasq-dns-89fc65cf9-zg284\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.424491 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.703096 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" event={"ID":"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507","Type":"ContainerStarted","Data":"cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.707638 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-cm5dq" event={"ID":"765efee7-b0c7-417f-8ad6-fee3d089bc50","Type":"ContainerStarted","Data":"b717f43396ad4516ea52f0f94c6eeab96522ec9934f263714841fd6374972f31"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.709849 4710 generic.go:334] "Generic (PLEG): container finished" podID="34948622-b486-47f0-bbb7-832f5ca6a1e7" containerID="9fdf852541c55aa2bc18ed6b6bedc735d9f85d14c120c89b2d22fbf99c473b63" exitCode=0 Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.709906 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a0e6-account-create-n8wcs" event={"ID":"34948622-b486-47f0-bbb7-832f5ca6a1e7","Type":"ContainerDied","Data":"9fdf852541c55aa2bc18ed6b6bedc735d9f85d14c120c89b2d22fbf99c473b63"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.709940 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a0e6-account-create-n8wcs" event={"ID":"34948622-b486-47f0-bbb7-832f5ca6a1e7","Type":"ContainerStarted","Data":"86d3267de855729cbe271faa002fd1edeb4cdc8a95203e8c5cdc18fa84ba251d"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.713621 4710 generic.go:334] "Generic (PLEG): container finished" podID="39dc6a17-1fa2-440c-9ee4-3b12755fc422" containerID="a36bbda7761dac810ffb248a657d9ec4e6e6e94fd02510553c6b3d9e12c84cde" exitCode=0 Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.713685 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-21d4-account-create-r2d77" event={"ID":"39dc6a17-1fa2-440c-9ee4-3b12755fc422","Type":"ContainerDied","Data":"a36bbda7761dac810ffb248a657d9ec4e6e6e94fd02510553c6b3d9e12c84cde"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.713712 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-21d4-account-create-r2d77" event={"ID":"39dc6a17-1fa2-440c-9ee4-3b12755fc422","Type":"ContainerStarted","Data":"ee2150342d9cb6619e46d218f3bf7b5d2dc370ca797f341f726a8073209033ae"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.737958 4710 generic.go:334] "Generic (PLEG): container finished" podID="03930039-3e8f-4a37-ae83-61f08a70aea4" containerID="9b3950de6e8766c1c2d5fe1bb14dfc026c167355de3b5946ddf9c8e01add197f" exitCode=0 Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.738603 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9aae-account-create-vt2hh" event={"ID":"03930039-3e8f-4a37-ae83-61f08a70aea4","Type":"ContainerDied","Data":"9b3950de6e8766c1c2d5fe1bb14dfc026c167355de3b5946ddf9c8e01add197f"} Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.738782 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9aae-account-create-vt2hh" event={"ID":"03930039-3e8f-4a37-ae83-61f08a70aea4","Type":"ContainerStarted","Data":"012b3657a2d7c4e78b99092d36893415b044bcb893e4140b185482f89773fb89"} Oct 02 13:07:02 crc kubenswrapper[4710]: W1002 13:07:02.946554 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64d612e0_1f54_4609_a642_bd78fd3a1269.slice/crio-6019d7290c6f4de18a7fb4b3f96da9e501d45d1b71cfa81d8ad934d66015066f WatchSource:0}: Error finding container 6019d7290c6f4de18a7fb4b3f96da9e501d45d1b71cfa81d8ad934d66015066f: Status 404 returned error can't find the container with id 6019d7290c6f4de18a7fb4b3f96da9e501d45d1b71cfa81d8ad934d66015066f Oct 02 13:07:02 crc kubenswrapper[4710]: I1002 13:07:02.951339 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89fc65cf9-zg284"] Oct 02 13:07:03 crc kubenswrapper[4710]: I1002 13:07:03.749482 4710 generic.go:334] "Generic (PLEG): container finished" podID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerID="ac46934e07d5ee7ef244faaf97821d55a665e7fd407e28f24aa9261894318bb4" exitCode=0 Oct 02 13:07:03 crc kubenswrapper[4710]: I1002 13:07:03.749575 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" event={"ID":"64d612e0-1f54-4609-a642-bd78fd3a1269","Type":"ContainerDied","Data":"ac46934e07d5ee7ef244faaf97821d55a665e7fd407e28f24aa9261894318bb4"} Oct 02 13:07:03 crc kubenswrapper[4710]: I1002 13:07:03.749641 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" event={"ID":"64d612e0-1f54-4609-a642-bd78fd3a1269","Type":"ContainerStarted","Data":"6019d7290c6f4de18a7fb4b3f96da9e501d45d1b71cfa81d8ad934d66015066f"} Oct 02 13:07:03 crc kubenswrapper[4710]: I1002 13:07:03.750358 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerName="dnsmasq-dns" containerID="cri-o://cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306" gracePeriod=10 Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.281236 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.440161 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.444896 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.451776 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.466948 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tt54\" (UniqueName: \"kubernetes.io/projected/34948622-b486-47f0-bbb7-832f5ca6a1e7-kube-api-access-5tt54\") pod \"34948622-b486-47f0-bbb7-832f5ca6a1e7\" (UID: \"34948622-b486-47f0-bbb7-832f5ca6a1e7\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.473457 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34948622-b486-47f0-bbb7-832f5ca6a1e7-kube-api-access-5tt54" (OuterVolumeSpecName: "kube-api-access-5tt54") pod "34948622-b486-47f0-bbb7-832f5ca6a1e7" (UID: "34948622-b486-47f0-bbb7-832f5ca6a1e7"). InnerVolumeSpecName "kube-api-access-5tt54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568296 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-nb\") pod \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568368 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-config\") pod \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568441 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66n2t\" (UniqueName: \"kubernetes.io/projected/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-kube-api-access-66n2t\") pod \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568499 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-swift-storage-0\") pod \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568605 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svf5f\" (UniqueName: \"kubernetes.io/projected/39dc6a17-1fa2-440c-9ee4-3b12755fc422-kube-api-access-svf5f\") pod \"39dc6a17-1fa2-440c-9ee4-3b12755fc422\" (UID: \"39dc6a17-1fa2-440c-9ee4-3b12755fc422\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568642 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcnsv\" (UniqueName: \"kubernetes.io/projected/03930039-3e8f-4a37-ae83-61f08a70aea4-kube-api-access-dcnsv\") pod \"03930039-3e8f-4a37-ae83-61f08a70aea4\" (UID: \"03930039-3e8f-4a37-ae83-61f08a70aea4\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568691 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-sb\") pod \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.568779 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-svc\") pod \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\" (UID: \"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507\") " Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.569182 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tt54\" (UniqueName: \"kubernetes.io/projected/34948622-b486-47f0-bbb7-832f5ca6a1e7-kube-api-access-5tt54\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.575979 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-kube-api-access-66n2t" (OuterVolumeSpecName: "kube-api-access-66n2t") pod "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" (UID: "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507"). InnerVolumeSpecName "kube-api-access-66n2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.602106 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39dc6a17-1fa2-440c-9ee4-3b12755fc422-kube-api-access-svf5f" (OuterVolumeSpecName: "kube-api-access-svf5f") pod "39dc6a17-1fa2-440c-9ee4-3b12755fc422" (UID: "39dc6a17-1fa2-440c-9ee4-3b12755fc422"). InnerVolumeSpecName "kube-api-access-svf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.609370 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03930039-3e8f-4a37-ae83-61f08a70aea4-kube-api-access-dcnsv" (OuterVolumeSpecName: "kube-api-access-dcnsv") pod "03930039-3e8f-4a37-ae83-61f08a70aea4" (UID: "03930039-3e8f-4a37-ae83-61f08a70aea4"). InnerVolumeSpecName "kube-api-access-dcnsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.671234 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svf5f\" (UniqueName: \"kubernetes.io/projected/39dc6a17-1fa2-440c-9ee4-3b12755fc422-kube-api-access-svf5f\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.671268 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcnsv\" (UniqueName: \"kubernetes.io/projected/03930039-3e8f-4a37-ae83-61f08a70aea4-kube-api-access-dcnsv\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.671280 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66n2t\" (UniqueName: \"kubernetes.io/projected/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-kube-api-access-66n2t\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.674322 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" (UID: "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.686137 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-config" (OuterVolumeSpecName: "config") pod "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" (UID: "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.696909 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" (UID: "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.702453 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" (UID: "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.710760 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" (UID: "f9e6d704-69b0-4e39-bed8-ef8ff6bc5507"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.760882 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9aae-account-create-vt2hh" event={"ID":"03930039-3e8f-4a37-ae83-61f08a70aea4","Type":"ContainerDied","Data":"012b3657a2d7c4e78b99092d36893415b044bcb893e4140b185482f89773fb89"} Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.760922 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="012b3657a2d7c4e78b99092d36893415b044bcb893e4140b185482f89773fb89" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.760916 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9aae-account-create-vt2hh" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.762810 4710 generic.go:334] "Generic (PLEG): container finished" podID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerID="cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306" exitCode=0 Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.762873 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" event={"ID":"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507","Type":"ContainerDied","Data":"cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306"} Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.762884 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.762899 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9c777fc7-mjjbd" event={"ID":"f9e6d704-69b0-4e39-bed8-ef8ff6bc5507","Type":"ContainerDied","Data":"63e27bf82ff788017f8ea9ece832018a26b0af1d8e6c93db08a8d5ed81f2bc61"} Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.762917 4710 scope.go:117] "RemoveContainer" containerID="cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.764060 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a0e6-account-create-n8wcs" event={"ID":"34948622-b486-47f0-bbb7-832f5ca6a1e7","Type":"ContainerDied","Data":"86d3267de855729cbe271faa002fd1edeb4cdc8a95203e8c5cdc18fa84ba251d"} Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.764083 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86d3267de855729cbe271faa002fd1edeb4cdc8a95203e8c5cdc18fa84ba251d" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.764091 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0e6-account-create-n8wcs" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.765445 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-21d4-account-create-r2d77" event={"ID":"39dc6a17-1fa2-440c-9ee4-3b12755fc422","Type":"ContainerDied","Data":"ee2150342d9cb6619e46d218f3bf7b5d2dc370ca797f341f726a8073209033ae"} Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.765476 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-21d4-account-create-r2d77" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.765481 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee2150342d9cb6619e46d218f3bf7b5d2dc370ca797f341f726a8073209033ae" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.766652 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" event={"ID":"64d612e0-1f54-4609-a642-bd78fd3a1269","Type":"ContainerStarted","Data":"d83566781e5002da12ac95f0b02ae2974a379a842b8ca9ba535984745afbb884"} Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.767503 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.772936 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.772971 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.772984 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.772996 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.773007 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.789815 4710 scope.go:117] "RemoveContainer" containerID="8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.851598 4710 scope.go:117] "RemoveContainer" containerID="cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306" Oct 02 13:07:04 crc kubenswrapper[4710]: E1002 13:07:04.853660 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306\": container with ID starting with cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306 not found: ID does not exist" containerID="cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.853735 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306"} err="failed to get container status \"cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306\": rpc error: code = NotFound desc = could not find container \"cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306\": container with ID starting with cfc2d1a335c0dc916734825df8c30af988dafd2b647d0582134c280f8f8cf306 not found: ID does not exist" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.854113 4710 scope.go:117] "RemoveContainer" containerID="8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a" Oct 02 13:07:04 crc kubenswrapper[4710]: E1002 13:07:04.854690 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a\": container with ID starting with 8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a not found: ID does not exist" containerID="8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.854740 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a"} err="failed to get container status \"8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a\": rpc error: code = NotFound desc = could not find container \"8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a\": container with ID starting with 8da2b4257d4ed30b9acd647c29edd591e8cfa17ce43419dac438a1f01366212a not found: ID does not exist" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.867302 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" podStartSLOduration=2.867256492 podStartE2EDuration="2.867256492s" podCreationTimestamp="2025-10-02 13:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:04.801350664 +0000 UTC m=+1068.907761567" watchObservedRunningTime="2025-10-02 13:07:04.867256492 +0000 UTC m=+1068.973667375" Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.870412 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d9c777fc7-mjjbd"] Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.883868 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d9c777fc7-mjjbd"] Oct 02 13:07:04 crc kubenswrapper[4710]: I1002 13:07:04.914000 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" path="/var/lib/kubelet/pods/f9e6d704-69b0-4e39-bed8-ef8ff6bc5507/volumes" Oct 02 13:07:07 crc kubenswrapper[4710]: I1002 13:07:07.794763 4710 generic.go:334] "Generic (PLEG): container finished" podID="765efee7-b0c7-417f-8ad6-fee3d089bc50" containerID="b717f43396ad4516ea52f0f94c6eeab96522ec9934f263714841fd6374972f31" exitCode=0 Oct 02 13:07:07 crc kubenswrapper[4710]: I1002 13:07:07.795450 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-cm5dq" event={"ID":"765efee7-b0c7-417f-8ad6-fee3d089bc50","Type":"ContainerDied","Data":"b717f43396ad4516ea52f0f94c6eeab96522ec9934f263714841fd6374972f31"} Oct 02 13:07:08 crc kubenswrapper[4710]: I1002 13:07:08.804960 4710 generic.go:334] "Generic (PLEG): container finished" podID="ad3b2f5f-f4d4-4a80-948d-275033ab9f05" containerID="23d65eded6f15a0f22f1963e12678f7c593c49010d2f67512979d07691a537a6" exitCode=0 Oct 02 13:07:08 crc kubenswrapper[4710]: I1002 13:07:08.805045 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf9wb" event={"ID":"ad3b2f5f-f4d4-4a80-948d-275033ab9f05","Type":"ContainerDied","Data":"23d65eded6f15a0f22f1963e12678f7c593c49010d2f67512979d07691a537a6"} Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.152973 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.248210 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-combined-ca-bundle\") pod \"765efee7-b0c7-417f-8ad6-fee3d089bc50\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.248268 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-db-sync-config-data\") pod \"765efee7-b0c7-417f-8ad6-fee3d089bc50\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.248327 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpbs7\" (UniqueName: \"kubernetes.io/projected/765efee7-b0c7-417f-8ad6-fee3d089bc50-kube-api-access-jpbs7\") pod \"765efee7-b0c7-417f-8ad6-fee3d089bc50\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.248353 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-config-data\") pod \"765efee7-b0c7-417f-8ad6-fee3d089bc50\" (UID: \"765efee7-b0c7-417f-8ad6-fee3d089bc50\") " Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.254050 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "765efee7-b0c7-417f-8ad6-fee3d089bc50" (UID: "765efee7-b0c7-417f-8ad6-fee3d089bc50"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.254333 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/765efee7-b0c7-417f-8ad6-fee3d089bc50-kube-api-access-jpbs7" (OuterVolumeSpecName: "kube-api-access-jpbs7") pod "765efee7-b0c7-417f-8ad6-fee3d089bc50" (UID: "765efee7-b0c7-417f-8ad6-fee3d089bc50"). InnerVolumeSpecName "kube-api-access-jpbs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.272158 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "765efee7-b0c7-417f-8ad6-fee3d089bc50" (UID: "765efee7-b0c7-417f-8ad6-fee3d089bc50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.295788 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-config-data" (OuterVolumeSpecName: "config-data") pod "765efee7-b0c7-417f-8ad6-fee3d089bc50" (UID: "765efee7-b0c7-417f-8ad6-fee3d089bc50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.350078 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.350112 4710 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.350126 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpbs7\" (UniqueName: \"kubernetes.io/projected/765efee7-b0c7-417f-8ad6-fee3d089bc50-kube-api-access-jpbs7\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.350141 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765efee7-b0c7-417f-8ad6-fee3d089bc50-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.815859 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-cm5dq" event={"ID":"765efee7-b0c7-417f-8ad6-fee3d089bc50","Type":"ContainerDied","Data":"00ce7925a7fc416b0244d667b9e958862aa127280db00db3278ec17d0c20df60"} Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.815908 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00ce7925a7fc416b0244d667b9e958862aa127280db00db3278ec17d0c20df60" Oct 02 13:07:09 crc kubenswrapper[4710]: I1002 13:07:09.815949 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-cm5dq" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.195539 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.265344 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-config-data\") pod \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.265525 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-combined-ca-bundle\") pod \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.265639 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jftj\" (UniqueName: \"kubernetes.io/projected/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-kube-api-access-7jftj\") pod \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\" (UID: \"ad3b2f5f-f4d4-4a80-948d-275033ab9f05\") " Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.275982 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-kube-api-access-7jftj" (OuterVolumeSpecName: "kube-api-access-7jftj") pod "ad3b2f5f-f4d4-4a80-948d-275033ab9f05" (UID: "ad3b2f5f-f4d4-4a80-948d-275033ab9f05"). InnerVolumeSpecName "kube-api-access-7jftj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.295671 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad3b2f5f-f4d4-4a80-948d-275033ab9f05" (UID: "ad3b2f5f-f4d4-4a80-948d-275033ab9f05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.318082 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-config-data" (OuterVolumeSpecName: "config-data") pod "ad3b2f5f-f4d4-4a80-948d-275033ab9f05" (UID: "ad3b2f5f-f4d4-4a80-948d-275033ab9f05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.367452 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jftj\" (UniqueName: \"kubernetes.io/projected/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-kube-api-access-7jftj\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.367496 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.367509 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3b2f5f-f4d4-4a80-948d-275033ab9f05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.826325 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf9wb" event={"ID":"ad3b2f5f-f4d4-4a80-948d-275033ab9f05","Type":"ContainerDied","Data":"287ab3ae101f86b43a0cca33cc6f5f4c7e02f3e87443b18ff38841208f1b672a"} Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.826396 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="287ab3ae101f86b43a0cca33cc6f5f4c7e02f3e87443b18ff38841208f1b672a" Oct 02 13:07:10 crc kubenswrapper[4710]: I1002 13:07:10.826426 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf9wb" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.144233 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89fc65cf9-zg284"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.144544 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="dnsmasq-dns" containerID="cri-o://d83566781e5002da12ac95f0b02ae2974a379a842b8ca9ba535984745afbb884" gracePeriod=10 Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.157573 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.198917 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f94b56449-g57dk"] Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199281 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3b2f5f-f4d4-4a80-948d-275033ab9f05" containerName="keystone-db-sync" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199292 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3b2f5f-f4d4-4a80-948d-275033ab9f05" containerName="keystone-db-sync" Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199304 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="765efee7-b0c7-417f-8ad6-fee3d089bc50" containerName="watcher-db-sync" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199310 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="765efee7-b0c7-417f-8ad6-fee3d089bc50" containerName="watcher-db-sync" Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199321 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03930039-3e8f-4a37-ae83-61f08a70aea4" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199327 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="03930039-3e8f-4a37-ae83-61f08a70aea4" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199338 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39dc6a17-1fa2-440c-9ee4-3b12755fc422" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199343 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="39dc6a17-1fa2-440c-9ee4-3b12755fc422" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199358 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerName="init" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199364 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerName="init" Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199369 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerName="dnsmasq-dns" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199375 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerName="dnsmasq-dns" Oct 02 13:07:11 crc kubenswrapper[4710]: E1002 13:07:11.199388 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34948622-b486-47f0-bbb7-832f5ca6a1e7" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199394 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="34948622-b486-47f0-bbb7-832f5ca6a1e7" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199554 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e6d704-69b0-4e39-bed8-ef8ff6bc5507" containerName="dnsmasq-dns" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199568 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="03930039-3e8f-4a37-ae83-61f08a70aea4" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199578 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="39dc6a17-1fa2-440c-9ee4-3b12755fc422" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199591 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3b2f5f-f4d4-4a80-948d-275033ab9f05" containerName="keystone-db-sync" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199604 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="34948622-b486-47f0-bbb7-832f5ca6a1e7" containerName="mariadb-account-create" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.199614 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="765efee7-b0c7-417f-8ad6-fee3d089bc50" containerName="watcher-db-sync" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.200491 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.256392 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-cf7wd"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.257567 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.274618 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.274982 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f94b56449-g57dk"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.282072 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p52kv" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.282584 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.282784 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.283839 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-sb\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.283891 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h72g7\" (UniqueName: \"kubernetes.io/projected/ec64e8d1-788a-4619-b936-830eab12aade-kube-api-access-h72g7\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.283944 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-swift-storage-0\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.283999 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-config\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.284020 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-svc\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.284041 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-nb\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.383287 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cf7wd"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.386724 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-combined-ca-bundle\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.401633 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-sb\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.401768 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-fernet-keys\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.401880 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h72g7\" (UniqueName: \"kubernetes.io/projected/ec64e8d1-788a-4619-b936-830eab12aade-kube-api-access-h72g7\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.401963 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-config-data\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.420956 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-credential-keys\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.421190 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-scripts\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.421288 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-swift-storage-0\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.422670 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-config\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.422803 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-svc\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.422897 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82xhw\" (UniqueName: \"kubernetes.io/projected/d9b58541-dcaa-47c4-b256-48ce93a3ab38-kube-api-access-82xhw\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.422978 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-nb\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.422576 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-swift-storage-0\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.410791 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-sb\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.424877 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-config\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.425490 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-svc\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.442288 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-nb\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.452650 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h72g7\" (UniqueName: \"kubernetes.io/projected/ec64e8d1-788a-4619-b936-830eab12aade-kube-api-access-h72g7\") pod \"dnsmasq-dns-7f94b56449-g57dk\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.517822 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.519305 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.525725 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-combined-ca-bundle\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.525981 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-fernet-keys\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.526120 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-config-data\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.526199 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-credential-keys\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.526261 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-scripts\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.526351 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82xhw\" (UniqueName: \"kubernetes.io/projected/d9b58541-dcaa-47c4-b256-48ce93a3ab38-kube-api-access-82xhw\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.532982 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-pvpl2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.533262 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.533883 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-combined-ca-bundle\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.541801 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.544160 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-fernet-keys\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.544909 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-config-data\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.559766 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-scripts\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.585403 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82xhw\" (UniqueName: \"kubernetes.io/projected/d9b58541-dcaa-47c4-b256-48ce93a3ab38-kube-api-access-82xhw\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.603949 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.628815 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.630688 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.632487 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-credential-keys\") pod \"keystone-bootstrap-cf7wd\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.635925 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq6jv\" (UniqueName: \"kubernetes.io/projected/8ecc2560-d8d5-405b-ae65-0112862e2171-kube-api-access-xq6jv\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.635997 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.636022 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ecc2560-d8d5-405b-ae65-0112862e2171-logs\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.636066 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.636090 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.636485 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.657898 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.704199 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.705678 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.708116 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.733959 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65856c9f97-ptvgn"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.735369 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738224 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skgrq\" (UniqueName: \"kubernetes.io/projected/3d51b0dc-7827-46ff-a400-98758053ef98-kube-api-access-skgrq\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738264 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq6jv\" (UniqueName: \"kubernetes.io/projected/8ecc2560-d8d5-405b-ae65-0112862e2171-kube-api-access-xq6jv\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738301 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738330 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738347 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ecc2560-d8d5-405b-ae65-0112862e2171-logs\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738376 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-config-data\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738398 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738415 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738434 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d51b0dc-7827-46ff-a400-98758053ef98-logs\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.738459 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.740414 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ecc2560-d8d5-405b-ae65-0112862e2171-logs\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.747629 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-tj68k" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.747942 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.748061 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.748153 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.748232 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.755523 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.757871 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.759589 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.797482 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq6jv\" (UniqueName: \"kubernetes.io/projected/8ecc2560-d8d5-405b-ae65-0112862e2171-kube-api-access-xq6jv\") pod \"watcher-decision-engine-0\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.802804 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dqkk2"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.803933 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.808294 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4rbpm" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.808468 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.808571 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.824866 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f94b56449-g57dk"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840698 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840763 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-config-data\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840788 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d887992a-ce0d-4e1c-926d-e1febbde1ca5-horizon-secret-key\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840815 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b868ba48-07b1-4a8e-8a4a-3c129840795c-logs\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840847 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csrlp\" (UniqueName: \"kubernetes.io/projected/b868ba48-07b1-4a8e-8a4a-3c129840795c-kube-api-access-csrlp\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840890 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skgrq\" (UniqueName: \"kubernetes.io/projected/3d51b0dc-7827-46ff-a400-98758053ef98-kube-api-access-skgrq\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840911 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btsqs\" (UniqueName: \"kubernetes.io/projected/d887992a-ce0d-4e1c-926d-e1febbde1ca5-kube-api-access-btsqs\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840935 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-scripts\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.840963 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841005 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-config-data\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841031 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-config-data\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841049 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d887992a-ce0d-4e1c-926d-e1febbde1ca5-logs\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841056 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65856c9f97-ptvgn"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841065 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d51b0dc-7827-46ff-a400-98758053ef98-logs\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841141 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.841401 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d51b0dc-7827-46ff-a400-98758053ef98-logs\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.852306 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.857272 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.859008 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-config-data\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.872871 4710 generic.go:334] "Generic (PLEG): container finished" podID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerID="d83566781e5002da12ac95f0b02ae2974a379a842b8ca9ba535984745afbb884" exitCode=0 Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.872916 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" event={"ID":"64d612e0-1f54-4609-a642-bd78fd3a1269","Type":"ContainerDied","Data":"d83566781e5002da12ac95f0b02ae2974a379a842b8ca9ba535984745afbb884"} Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.895788 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dqkk2"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.911200 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skgrq\" (UniqueName: \"kubernetes.io/projected/3d51b0dc-7827-46ff-a400-98758053ef98-kube-api-access-skgrq\") pod \"watcher-api-0\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " pod="openstack/watcher-api-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.914856 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.921829 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4m9cc"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.922954 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.930474 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.930653 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.930765 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-c7scg" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942669 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-config-data\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942724 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btsqs\" (UniqueName: \"kubernetes.io/projected/d887992a-ce0d-4e1c-926d-e1febbde1ca5-kube-api-access-btsqs\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942773 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-scripts\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942855 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6216b148-2847-437c-8437-1f574973fb87-logs\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942893 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kssg4\" (UniqueName: \"kubernetes.io/projected/6216b148-2847-437c-8437-1f574973fb87-kube-api-access-kssg4\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942944 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-config-data\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.942969 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d887992a-ce0d-4e1c-926d-e1febbde1ca5-logs\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943021 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943049 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-scripts\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943078 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-config-data\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943109 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d887992a-ce0d-4e1c-926d-e1febbde1ca5-horizon-secret-key\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943139 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b868ba48-07b1-4a8e-8a4a-3c129840795c-logs\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943162 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csrlp\" (UniqueName: \"kubernetes.io/projected/b868ba48-07b1-4a8e-8a4a-3c129840795c-kube-api-access-csrlp\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.943192 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-combined-ca-bundle\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.944328 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b868ba48-07b1-4a8e-8a4a-3c129840795c-logs\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.944358 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d887992a-ce0d-4e1c-926d-e1febbde1ca5-logs\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.944831 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-scripts\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.945600 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-config-data\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.950690 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-config-data\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.957206 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d887992a-ce0d-4e1c-926d-e1febbde1ca5-horizon-secret-key\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.962023 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-swkp8"] Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.966592 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.983229 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.989291 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-458k4" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.989478 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 13:07:11 crc kubenswrapper[4710]: I1002 13:07:11.989587 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.004147 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csrlp\" (UniqueName: \"kubernetes.io/projected/b868ba48-07b1-4a8e-8a4a-3c129840795c-kube-api-access-csrlp\") pod \"watcher-applier-0\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " pod="openstack/watcher-applier-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.020208 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4m9cc"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.034538 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btsqs\" (UniqueName: \"kubernetes.io/projected/d887992a-ce0d-4e1c-926d-e1febbde1ca5-kube-api-access-btsqs\") pod \"horizon-65856c9f97-ptvgn\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.044726 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-scripts\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.044798 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kssg4\" (UniqueName: \"kubernetes.io/projected/6216b148-2847-437c-8437-1f574973fb87-kube-api-access-kssg4\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.044871 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-combined-ca-bundle\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.044900 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-scripts\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.044938 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrzkv\" (UniqueName: \"kubernetes.io/projected/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-kube-api-access-hrzkv\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045007 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-config\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045027 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-combined-ca-bundle\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045086 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq4q5\" (UniqueName: \"kubernetes.io/projected/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-kube-api-access-fq4q5\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045118 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-config-data\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045163 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-config-data\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045198 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-db-sync-config-data\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045215 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-etc-machine-id\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045249 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6216b148-2847-437c-8437-1f574973fb87-logs\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.045267 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-combined-ca-bundle\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.046506 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6216b148-2847-437c-8437-1f574973fb87-logs\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.049142 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-config-data\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.057806 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-combined-ca-bundle\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.061218 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-849796779c-f5465"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.062908 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.080972 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-swkp8"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.081313 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-scripts\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.091646 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.091691 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kssg4\" (UniqueName: \"kubernetes.io/projected/6216b148-2847-437c-8437-1f574973fb87-kube-api-access-kssg4\") pod \"placement-db-sync-dqkk2\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.095980 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849796779c-f5465"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.098621 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.104494 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58dc9c68c-xp6ns"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.107093 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.118840 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-nwcvq"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.120185 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.123464 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gq2s5" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.123680 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.136158 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58dc9c68c-xp6ns"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.136524 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.143279 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147617 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bjj\" (UniqueName: \"kubernetes.io/projected/134d8dfd-37b4-4d67-94d7-aa377e780bae-kube-api-access-99bjj\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147672 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-combined-ca-bundle\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147696 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-nb\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147722 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrzkv\" (UniqueName: \"kubernetes.io/projected/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-kube-api-access-hrzkv\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147767 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfr8t\" (UniqueName: \"kubernetes.io/projected/6d28154a-1fc9-4d45-bb67-ecac912c5d01-kube-api-access-zfr8t\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147791 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-config\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147809 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d8dfd-37b4-4d67-94d7-aa377e780bae-logs\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147843 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq4q5\" (UniqueName: \"kubernetes.io/projected/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-kube-api-access-fq4q5\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147862 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/134d8dfd-37b4-4d67-94d7-aa377e780bae-horizon-secret-key\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147882 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-swift-storage-0\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147901 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-config-data\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147926 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-svc\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147948 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-config-data\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147970 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-db-sync-config-data\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.147987 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-etc-machine-id\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.148007 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-config\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.148023 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-scripts\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.148059 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-combined-ca-bundle\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.148087 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-scripts\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.148122 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-sb\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.149483 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-etc-machine-id\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.158890 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-scripts\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.161578 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.165192 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-combined-ca-bundle\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.165216 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-combined-ca-bundle\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.167978 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-config-data\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.169640 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-config\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.171047 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nwcvq"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.171759 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-db-sync-config-data\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.172232 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrzkv\" (UniqueName: \"kubernetes.io/projected/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-kube-api-access-hrzkv\") pod \"cinder-db-sync-4m9cc\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.179002 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq4q5\" (UniqueName: \"kubernetes.io/projected/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-kube-api-access-fq4q5\") pod \"neutron-db-sync-swkp8\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.205343 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.233135 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.233855 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.247509 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.249010 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269617 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/134d8dfd-37b4-4d67-94d7-aa377e780bae-horizon-secret-key\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269661 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-swift-storage-0\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269701 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-config-data\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269775 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-db-sync-config-data\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269815 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-svc\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269910 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-config\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.269957 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-scripts\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.270054 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-sb\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.275735 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/134d8dfd-37b4-4d67-94d7-aa377e780bae-horizon-secret-key\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.281119 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-swift-storage-0\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.282106 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-sb\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.282226 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-scripts\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.282515 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bjj\" (UniqueName: \"kubernetes.io/projected/134d8dfd-37b4-4d67-94d7-aa377e780bae-kube-api-access-99bjj\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.282568 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-nb\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.283234 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-config\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.282636 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkj2j\" (UniqueName: \"kubernetes.io/projected/f53ffe3d-d27a-41dd-a98b-4debca318ec4-kube-api-access-pkj2j\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.284171 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-svc\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.284298 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfr8t\" (UniqueName: \"kubernetes.io/projected/6d28154a-1fc9-4d45-bb67-ecac912c5d01-kube-api-access-zfr8t\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.284375 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d8dfd-37b4-4d67-94d7-aa377e780bae-logs\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.284492 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-combined-ca-bundle\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.284300 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-config-data\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.285576 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-nb\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.285607 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d8dfd-37b4-4d67-94d7-aa377e780bae-logs\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.286313 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.313295 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bjj\" (UniqueName: \"kubernetes.io/projected/134d8dfd-37b4-4d67-94d7-aa377e780bae-kube-api-access-99bjj\") pod \"horizon-58dc9c68c-xp6ns\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.313343 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.314821 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.317132 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.317297 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.317409 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6fkzg" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.317510 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.323425 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfr8t\" (UniqueName: \"kubernetes.io/projected/6d28154a-1fc9-4d45-bb67-ecac912c5d01-kube-api-access-zfr8t\") pod \"dnsmasq-dns-849796779c-f5465\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.323501 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.330684 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387262 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-config-data\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387327 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-log-httpd\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387355 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-config-data\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387417 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkj2j\" (UniqueName: \"kubernetes.io/projected/f53ffe3d-d27a-41dd-a98b-4debca318ec4-kube-api-access-pkj2j\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387432 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387449 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-254sx\" (UniqueName: \"kubernetes.io/projected/9e4fe495-9150-420c-8d44-e56e8808b048-kube-api-access-254sx\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387477 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387507 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-logs\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387538 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-scripts\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387557 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387579 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-combined-ca-bundle\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387604 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387635 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-scripts\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387657 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crqtp\" (UniqueName: \"kubernetes.io/projected/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-kube-api-access-crqtp\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387677 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-db-sync-config-data\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387737 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-run-httpd\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387809 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.387843 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.395006 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-combined-ca-bundle\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.399035 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-db-sync-config-data\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.405332 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.414819 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkj2j\" (UniqueName: \"kubernetes.io/projected/f53ffe3d-d27a-41dd-a98b-4debca318ec4-kube-api-access-pkj2j\") pod \"barbican-db-sync-nwcvq\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.447192 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.468598 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.494926 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-config-data\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495676 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-log-httpd\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495705 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-config-data\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495773 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495790 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-254sx\" (UniqueName: \"kubernetes.io/projected/9e4fe495-9150-420c-8d44-e56e8808b048-kube-api-access-254sx\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495826 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495871 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-logs\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495909 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-scripts\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495930 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495963 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.495993 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-scripts\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.498018 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crqtp\" (UniqueName: \"kubernetes.io/projected/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-kube-api-access-crqtp\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.498119 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-run-httpd\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.498149 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.498183 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.499001 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-run-httpd\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.499603 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.501662 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-logs\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.502133 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.509104 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-log-httpd\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.509672 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.511521 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.513702 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-config-data\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.515785 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-scripts\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.516350 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-scripts\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.521565 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.522904 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crqtp\" (UniqueName: \"kubernetes.io/projected/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-kube-api-access-crqtp\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.535143 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-254sx\" (UniqueName: \"kubernetes.io/projected/9e4fe495-9150-420c-8d44-e56e8808b048-kube-api-access-254sx\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.552408 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.554678 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-config-data\") pod \"ceilometer-0\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.567274 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.610125 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.661027 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.667108 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f94b56449-g57dk"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.837645 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.841541 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.844651 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.844928 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.852416 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939691 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939771 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzjxp\" (UniqueName: \"kubernetes.io/projected/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-kube-api-access-gzjxp\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939799 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939857 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939873 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939915 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939941 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.939964 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-logs\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.973282 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" event={"ID":"ec64e8d1-788a-4619-b936-830eab12aade","Type":"ContainerStarted","Data":"c901cb247593c3ec9bff07197bdb30de2f3964db181163cefbe9f5315160c5bd"} Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.973318 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" event={"ID":"64d612e0-1f54-4609-a642-bd78fd3a1269","Type":"ContainerDied","Data":"6019d7290c6f4de18a7fb4b3f96da9e501d45d1b71cfa81d8ad934d66015066f"} Oct 02 13:07:12 crc kubenswrapper[4710]: I1002 13:07:12.973330 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6019d7290c6f4de18a7fb4b3f96da9e501d45d1b71cfa81d8ad934d66015066f" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.036582 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.041820 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.041924 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.041947 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.042005 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.046219 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.042042 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.049187 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-logs\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.049392 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.049448 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzjxp\" (UniqueName: \"kubernetes.io/projected/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-kube-api-access-gzjxp\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.054609 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-logs\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.055173 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.064063 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.075625 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzjxp\" (UniqueName: \"kubernetes.io/projected/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-kube-api-access-gzjxp\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.098085 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.116358 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.131294 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.146195 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.150108 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wdld\" (UniqueName: \"kubernetes.io/projected/64d612e0-1f54-4609-a642-bd78fd3a1269-kube-api-access-9wdld\") pod \"64d612e0-1f54-4609-a642-bd78fd3a1269\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.150233 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-sb\") pod \"64d612e0-1f54-4609-a642-bd78fd3a1269\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.150275 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-svc\") pod \"64d612e0-1f54-4609-a642-bd78fd3a1269\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.150291 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-config\") pod \"64d612e0-1f54-4609-a642-bd78fd3a1269\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.150361 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-swift-storage-0\") pod \"64d612e0-1f54-4609-a642-bd78fd3a1269\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.150482 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-nb\") pod \"64d612e0-1f54-4609-a642-bd78fd3a1269\" (UID: \"64d612e0-1f54-4609-a642-bd78fd3a1269\") " Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.168396 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64d612e0-1f54-4609-a642-bd78fd3a1269-kube-api-access-9wdld" (OuterVolumeSpecName: "kube-api-access-9wdld") pod "64d612e0-1f54-4609-a642-bd78fd3a1269" (UID: "64d612e0-1f54-4609-a642-bd78fd3a1269"). InnerVolumeSpecName "kube-api-access-9wdld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.205655 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cf7wd"] Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.214786 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:07:13 crc kubenswrapper[4710]: W1002 13:07:13.240769 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9b58541_dcaa_47c4_b256_48ce93a3ab38.slice/crio-74c8fc1c2fd9efd77b0cea27cc77dc5ad5db5a137cb1cada288e59fecc34a778 WatchSource:0}: Error finding container 74c8fc1c2fd9efd77b0cea27cc77dc5ad5db5a137cb1cada288e59fecc34a778: Status 404 returned error can't find the container with id 74c8fc1c2fd9efd77b0cea27cc77dc5ad5db5a137cb1cada288e59fecc34a778 Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.259756 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wdld\" (UniqueName: \"kubernetes.io/projected/64d612e0-1f54-4609-a642-bd78fd3a1269-kube-api-access-9wdld\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.263435 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "64d612e0-1f54-4609-a642-bd78fd3a1269" (UID: "64d612e0-1f54-4609-a642-bd78fd3a1269"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.264444 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-config" (OuterVolumeSpecName: "config") pod "64d612e0-1f54-4609-a642-bd78fd3a1269" (UID: "64d612e0-1f54-4609-a642-bd78fd3a1269"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.266923 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "64d612e0-1f54-4609-a642-bd78fd3a1269" (UID: "64d612e0-1f54-4609-a642-bd78fd3a1269"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.276075 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "64d612e0-1f54-4609-a642-bd78fd3a1269" (UID: "64d612e0-1f54-4609-a642-bd78fd3a1269"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.281038 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64d612e0-1f54-4609-a642-bd78fd3a1269" (UID: "64d612e0-1f54-4609-a642-bd78fd3a1269"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.346210 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dqkk2"] Oct 02 13:07:13 crc kubenswrapper[4710]: W1002 13:07:13.349427 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6216b148_2847_437c_8437_1f574973fb87.slice/crio-9020a48a490dd2d83a5762bc2434ba1b9b3e521cdbae6c9dae39aa7da8719908 WatchSource:0}: Error finding container 9020a48a490dd2d83a5762bc2434ba1b9b3e521cdbae6c9dae39aa7da8719908: Status 404 returned error can't find the container with id 9020a48a490dd2d83a5762bc2434ba1b9b3e521cdbae6c9dae39aa7da8719908 Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.360443 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.360979 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.361002 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.361012 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.361020 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.361028 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64d612e0-1f54-4609-a642-bd78fd3a1269-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.389639 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:13 crc kubenswrapper[4710]: W1002 13:07:13.572305 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978ba8dd_e49f_40cc_a4f9_4dd5658e4001.slice/crio-3853388b578afd4b9961717af12dc32334e02a6f0f5912b901769ca93fa8a0c5 WatchSource:0}: Error finding container 3853388b578afd4b9961717af12dc32334e02a6f0f5912b901769ca93fa8a0c5: Status 404 returned error can't find the container with id 3853388b578afd4b9961717af12dc32334e02a6f0f5912b901769ca93fa8a0c5 Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.582670 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.611344 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-swkp8"] Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.653433 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nwcvq"] Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.950078 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dqkk2" event={"ID":"6216b148-2847-437c-8437-1f574973fb87","Type":"ContainerStarted","Data":"9020a48a490dd2d83a5762bc2434ba1b9b3e521cdbae6c9dae39aa7da8719908"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.955996 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerStarted","Data":"f891f4584ab0f0269dea4e6ebcf7c7e1008c21c3294c9d14603c5ca9b9e2d3f8"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.963878 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-swkp8" event={"ID":"978ba8dd-e49f-40cc-a4f9-4dd5658e4001","Type":"ContainerStarted","Data":"258fdf3876f9e30b2aeb09f882a559bc54129f5529eed0bc6e82425ed444a55a"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.963929 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-swkp8" event={"ID":"978ba8dd-e49f-40cc-a4f9-4dd5658e4001","Type":"ContainerStarted","Data":"3853388b578afd4b9961717af12dc32334e02a6f0f5912b901769ca93fa8a0c5"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.972124 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d51b0dc-7827-46ff-a400-98758053ef98","Type":"ContainerStarted","Data":"5a44308221361954713d3579f7f1efa3d323e878ee244751010bc754ede83449"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.972169 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d51b0dc-7827-46ff-a400-98758053ef98","Type":"ContainerStarted","Data":"1df5a1f636a7627aa4b90bcf0427e762efec93031ab7ab40b01b2e0e729136cc"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.975094 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"b868ba48-07b1-4a8e-8a4a-3c129840795c","Type":"ContainerStarted","Data":"d8533f893ed916d916c08d9eb7ae19d2a5a85d67322ed63419519da0f68c2ae1"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.976946 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cf7wd" event={"ID":"d9b58541-dcaa-47c4-b256-48ce93a3ab38","Type":"ContainerStarted","Data":"56c067acd68f3ffaa2c02e6108b6721175d00d25a2440ada8b051717cc8a236a"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.976982 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cf7wd" event={"ID":"d9b58541-dcaa-47c4-b256-48ce93a3ab38","Type":"ContainerStarted","Data":"74c8fc1c2fd9efd77b0cea27cc77dc5ad5db5a137cb1cada288e59fecc34a778"} Oct 02 13:07:13 crc kubenswrapper[4710]: I1002 13:07:13.982913 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nwcvq" event={"ID":"f53ffe3d-d27a-41dd-a98b-4debca318ec4","Type":"ContainerStarted","Data":"1f7fdfba007e7488b49e275a0c9a3ad53eaca5a6c22e531e45bfa32e377b3ca6"} Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:13.990323 4710 generic.go:334] "Generic (PLEG): container finished" podID="ec64e8d1-788a-4619-b936-830eab12aade" containerID="853080f018fe65d25a9a955dabe5b172bb749db9190b92444cfe6a55cc41d6c9" exitCode=0 Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:13.991211 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" event={"ID":"ec64e8d1-788a-4619-b936-830eab12aade","Type":"ContainerDied","Data":"853080f018fe65d25a9a955dabe5b172bb749db9190b92444cfe6a55cc41d6c9"} Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:13.991939 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.004209 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849796779c-f5465"] Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.008009 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-swkp8" podStartSLOduration=3.007989286 podStartE2EDuration="3.007989286s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:13.991551372 +0000 UTC m=+1078.097962255" watchObservedRunningTime="2025-10-02 13:07:14.007989286 +0000 UTC m=+1078.114400169" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.075971 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:07:14 crc kubenswrapper[4710]: W1002 13:07:14.103553 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod134d8dfd_37b4_4d67_94d7_aa377e780bae.slice/crio-baec0231e1b89afa894a6a06b2f22dee2bee98108c228e89ce12da464bd8cbc1 WatchSource:0}: Error finding container baec0231e1b89afa894a6a06b2f22dee2bee98108c228e89ce12da464bd8cbc1: Status 404 returned error can't find the container with id baec0231e1b89afa894a6a06b2f22dee2bee98108c228e89ce12da464bd8cbc1 Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.133159 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58dc9c68c-xp6ns"] Oct 02 13:07:14 crc kubenswrapper[4710]: W1002 13:07:14.168923 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd887992a_ce0d_4e1c_926d_e1febbde1ca5.slice/crio-c99b4a33af205726d0a02c88c3152750e4424deb5864667d42e07e1d1199996b WatchSource:0}: Error finding container c99b4a33af205726d0a02c88c3152750e4424deb5864667d42e07e1d1199996b: Status 404 returned error can't find the container with id c99b4a33af205726d0a02c88c3152750e4424deb5864667d42e07e1d1199996b Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.187660 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4m9cc"] Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.192789 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-cf7wd" podStartSLOduration=3.192766747 podStartE2EDuration="3.192766747s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:14.090283277 +0000 UTC m=+1078.196694160" watchObservedRunningTime="2025-10-02 13:07:14.192766747 +0000 UTC m=+1078.299177630" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.224932 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65856c9f97-ptvgn"] Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.231266 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89fc65cf9-zg284"] Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.242499 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89fc65cf9-zg284"] Oct 02 13:07:14 crc kubenswrapper[4710]: W1002 13:07:14.245233 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e4fe495_9150_420c_8d44_e56e8808b048.slice/crio-71316dff5f4aa8d4ebbb8ea92b88ad74f1c5892e8fc7fb32bf906cd4408320da WatchSource:0}: Error finding container 71316dff5f4aa8d4ebbb8ea92b88ad74f1c5892e8fc7fb32bf906cd4408320da: Status 404 returned error can't find the container with id 71316dff5f4aa8d4ebbb8ea92b88ad74f1c5892e8fc7fb32bf906cd4408320da Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.252801 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.716056 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.804087 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h72g7\" (UniqueName: \"kubernetes.io/projected/ec64e8d1-788a-4619-b936-830eab12aade-kube-api-access-h72g7\") pod \"ec64e8d1-788a-4619-b936-830eab12aade\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.804139 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-config\") pod \"ec64e8d1-788a-4619-b936-830eab12aade\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.804254 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-sb\") pod \"ec64e8d1-788a-4619-b936-830eab12aade\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.804296 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-svc\") pod \"ec64e8d1-788a-4619-b936-830eab12aade\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.804453 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-swift-storage-0\") pod \"ec64e8d1-788a-4619-b936-830eab12aade\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.804475 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-nb\") pod \"ec64e8d1-788a-4619-b936-830eab12aade\" (UID: \"ec64e8d1-788a-4619-b936-830eab12aade\") " Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.823856 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec64e8d1-788a-4619-b936-830eab12aade-kube-api-access-h72g7" (OuterVolumeSpecName: "kube-api-access-h72g7") pod "ec64e8d1-788a-4619-b936-830eab12aade" (UID: "ec64e8d1-788a-4619-b936-830eab12aade"). InnerVolumeSpecName "kube-api-access-h72g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.833962 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ec64e8d1-788a-4619-b936-830eab12aade" (UID: "ec64e8d1-788a-4619-b936-830eab12aade"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.846663 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ec64e8d1-788a-4619-b936-830eab12aade" (UID: "ec64e8d1-788a-4619-b936-830eab12aade"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.847408 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec64e8d1-788a-4619-b936-830eab12aade" (UID: "ec64e8d1-788a-4619-b936-830eab12aade"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.854354 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ec64e8d1-788a-4619-b936-830eab12aade" (UID: "ec64e8d1-788a-4619-b936-830eab12aade"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.878664 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-config" (OuterVolumeSpecName: "config") pod "ec64e8d1-788a-4619-b936-830eab12aade" (UID: "ec64e8d1-788a-4619-b936-830eab12aade"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.915042 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.915082 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.915094 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.915202 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h72g7\" (UniqueName: \"kubernetes.io/projected/ec64e8d1-788a-4619-b936-830eab12aade-kube-api-access-h72g7\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.915213 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.915284 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec64e8d1-788a-4619-b936-830eab12aade-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.941916 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" path="/var/lib/kubelet/pods/64d612e0-1f54-4609-a642-bd78fd3a1269/volumes" Oct 02 13:07:14 crc kubenswrapper[4710]: I1002 13:07:14.956276 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.023988 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9e4fe495-9150-420c-8d44-e56e8808b048","Type":"ContainerStarted","Data":"71316dff5f4aa8d4ebbb8ea92b88ad74f1c5892e8fc7fb32bf906cd4408320da"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.032175 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.032163 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f94b56449-g57dk" event={"ID":"ec64e8d1-788a-4619-b936-830eab12aade","Type":"ContainerDied","Data":"c901cb247593c3ec9bff07197bdb30de2f3964db181163cefbe9f5315160c5bd"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.032322 4710 scope.go:117] "RemoveContainer" containerID="853080f018fe65d25a9a955dabe5b172bb749db9190b92444cfe6a55cc41d6c9" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.037563 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerStarted","Data":"2d484683f04090875bc64dae71a5cd9d7078a47a72bc1956266bcc3edf2b6691"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.041316 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dc9c68c-xp6ns" event={"ID":"134d8dfd-37b4-4d67-94d7-aa377e780bae","Type":"ContainerStarted","Data":"baec0231e1b89afa894a6a06b2f22dee2bee98108c228e89ce12da464bd8cbc1"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.044201 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d51b0dc-7827-46ff-a400-98758053ef98","Type":"ContainerStarted","Data":"05a23d462401444d029bcda58f52f3b6e480136483d05d729efb4121f5175d26"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.045421 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.050472 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m9cc" event={"ID":"6152e5af-f395-4bca-9aa1-19d17ec0ffa2","Type":"ContainerStarted","Data":"3fe04575f67980db575ed297ed21e8ea7202bb48311456d9f337e4358e02f285"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.057464 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65856c9f97-ptvgn" event={"ID":"d887992a-ce0d-4e1c-926d-e1febbde1ca5","Type":"ContainerStarted","Data":"c99b4a33af205726d0a02c88c3152750e4424deb5864667d42e07e1d1199996b"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.125470 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f94b56449-g57dk"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.144978 4710 generic.go:334] "Generic (PLEG): container finished" podID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerID="ddfaea0fa96c0d46a5320c401e70052029af0e813a774b8515bc1dfc32ec0340" exitCode=0 Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.145419 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849796779c-f5465" event={"ID":"6d28154a-1fc9-4d45-bb67-ecac912c5d01","Type":"ContainerDied","Data":"ddfaea0fa96c0d46a5320c401e70052029af0e813a774b8515bc1dfc32ec0340"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.145501 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849796779c-f5465" event={"ID":"6d28154a-1fc9-4d45-bb67-ecac912c5d01","Type":"ContainerStarted","Data":"7d2bed6ee2ae2b48b30fc3bef8960567e1132db5820fbad777c2fd0d7084b629"} Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.226641 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f94b56449-g57dk"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.226816 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=4.226795344 podStartE2EDuration="4.226795344s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:15.111707417 +0000 UTC m=+1079.218118300" watchObservedRunningTime="2025-10-02 13:07:15.226795344 +0000 UTC m=+1079.333206227" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.382209 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.395740 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.415162 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58dc9c68c-xp6ns"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.456392 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.505695 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-588846bfb9-5mhkc"] Oct 02 13:07:15 crc kubenswrapper[4710]: E1002 13:07:15.510002 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec64e8d1-788a-4619-b936-830eab12aade" containerName="init" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.510181 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec64e8d1-788a-4619-b936-830eab12aade" containerName="init" Oct 02 13:07:15 crc kubenswrapper[4710]: E1002 13:07:15.510245 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="dnsmasq-dns" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.510304 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="dnsmasq-dns" Oct 02 13:07:15 crc kubenswrapper[4710]: E1002 13:07:15.510369 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="init" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.510416 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="init" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.510671 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec64e8d1-788a-4619-b936-830eab12aade" containerName="init" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.510734 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="dnsmasq-dns" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.511876 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.546168 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-588846bfb9-5mhkc"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.663717 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.676843 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-config-data\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.676914 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rtkd\" (UniqueName: \"kubernetes.io/projected/6b39391f-b6e4-4275-941b-63484ea90c69-kube-api-access-8rtkd\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.676937 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-scripts\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.677049 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b39391f-b6e4-4275-941b-63484ea90c69-logs\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.677074 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b39391f-b6e4-4275-941b-63484ea90c69-horizon-secret-key\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.778626 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b39391f-b6e4-4275-941b-63484ea90c69-horizon-secret-key\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.778697 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-config-data\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.778765 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rtkd\" (UniqueName: \"kubernetes.io/projected/6b39391f-b6e4-4275-941b-63484ea90c69-kube-api-access-8rtkd\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.778792 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-scripts\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.778971 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b39391f-b6e4-4275-941b-63484ea90c69-logs\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.779447 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b39391f-b6e4-4275-941b-63484ea90c69-logs\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.781576 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-config-data\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.782921 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-scripts\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.792251 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b39391f-b6e4-4275-941b-63484ea90c69-horizon-secret-key\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.825449 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rtkd\" (UniqueName: \"kubernetes.io/projected/6b39391f-b6e4-4275-941b-63484ea90c69-kube-api-access-8rtkd\") pod \"horizon-588846bfb9-5mhkc\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:15 crc kubenswrapper[4710]: I1002 13:07:15.850548 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:16 crc kubenswrapper[4710]: I1002 13:07:16.931649 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec64e8d1-788a-4619-b936-830eab12aade" path="/var/lib/kubelet/pods/ec64e8d1-788a-4619-b936-830eab12aade/volumes" Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.101387 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.219169 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.219314 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api-log" containerID="cri-o://5a44308221361954713d3579f7f1efa3d323e878ee244751010bc754ede83449" gracePeriod=30 Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.219604 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9e4fe495-9150-420c-8d44-e56e8808b048","Type":"ContainerStarted","Data":"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f"} Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.219887 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" containerID="cri-o://05a23d462401444d029bcda58f52f3b6e480136483d05d729efb4121f5175d26" gracePeriod=30 Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.237607 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.151:9322/\": EOF" Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.237701 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.151:9322/\": EOF" Oct 02 13:07:17 crc kubenswrapper[4710]: W1002 13:07:17.339890 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1edb453e_c0a3_4639_96fc_d78a7c9c9e0a.slice/crio-e05ea0d02c95783057d7aa844d6b5bc208cc9799f31e8c1f4dc1f653fcbdf88d WatchSource:0}: Error finding container e05ea0d02c95783057d7aa844d6b5bc208cc9799f31e8c1f4dc1f653fcbdf88d: Status 404 returned error can't find the container with id e05ea0d02c95783057d7aa844d6b5bc208cc9799f31e8c1f4dc1f653fcbdf88d Oct 02 13:07:17 crc kubenswrapper[4710]: I1002 13:07:17.427153 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89fc65cf9-zg284" podUID="64d612e0-1f54-4609-a642-bd78fd3a1269" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: i/o timeout" Oct 02 13:07:18 crc kubenswrapper[4710]: I1002 13:07:18.231355 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a","Type":"ContainerStarted","Data":"e05ea0d02c95783057d7aa844d6b5bc208cc9799f31e8c1f4dc1f653fcbdf88d"} Oct 02 13:07:18 crc kubenswrapper[4710]: I1002 13:07:18.234282 4710 generic.go:334] "Generic (PLEG): container finished" podID="3d51b0dc-7827-46ff-a400-98758053ef98" containerID="5a44308221361954713d3579f7f1efa3d323e878ee244751010bc754ede83449" exitCode=143 Oct 02 13:07:18 crc kubenswrapper[4710]: I1002 13:07:18.234357 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d51b0dc-7827-46ff-a400-98758053ef98","Type":"ContainerDied","Data":"5a44308221361954713d3579f7f1efa3d323e878ee244751010bc754ede83449"} Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.352812 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.151:9322/\": read tcp 10.217.0.2:44470->10.217.0.151:9322: read: connection reset by peer" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.354036 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.151:9322/\": dial tcp 10.217.0.151:9322: connect: connection refused" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.759401 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65856c9f97-ptvgn"] Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.786060 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f85867556-bdslh"] Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.791399 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.800841 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.821627 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f85867556-bdslh"] Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.889560 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-588846bfb9-5mhkc"] Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.953716 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-secret-key\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.964875 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-tls-certs\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.965530 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-config-data\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.965634 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-scripts\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.967655 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-combined-ca-bundle\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.967697 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9krh2\" (UniqueName: \"kubernetes.io/projected/00020afa-1e01-4a2c-924e-2aa55bac0e85-kube-api-access-9krh2\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.967894 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00020afa-1e01-4a2c-924e-2aa55bac0e85-logs\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.989317 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58474b8ddd-nzgx9"] Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.990847 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58474b8ddd-nzgx9"] Oct 02 13:07:20 crc kubenswrapper[4710]: I1002 13:07:20.990955 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069262 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-tls-certs\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069363 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-config-data\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069398 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-scripts\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069414 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-combined-ca-bundle\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069429 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9krh2\" (UniqueName: \"kubernetes.io/projected/00020afa-1e01-4a2c-924e-2aa55bac0e85-kube-api-access-9krh2\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069463 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00020afa-1e01-4a2c-924e-2aa55bac0e85-logs\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.069501 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-secret-key\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.070521 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-scripts\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.071443 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00020afa-1e01-4a2c-924e-2aa55bac0e85-logs\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.072166 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-config-data\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.076123 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-secret-key\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.077604 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-tls-certs\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.089348 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-combined-ca-bundle\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.099058 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9krh2\" (UniqueName: \"kubernetes.io/projected/00020afa-1e01-4a2c-924e-2aa55bac0e85-kube-api-access-9krh2\") pod \"horizon-5f85867556-bdslh\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.134480 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171149 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8267ea9-4756-45d5-a7bc-8985275cf297-config-data\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171230 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8267ea9-4756-45d5-a7bc-8985275cf297-logs\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171268 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgcll\" (UniqueName: \"kubernetes.io/projected/b8267ea9-4756-45d5-a7bc-8985275cf297-kube-api-access-bgcll\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171288 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-combined-ca-bundle\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171404 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-horizon-secret-key\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171430 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8267ea9-4756-45d5-a7bc-8985275cf297-scripts\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.171465 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-horizon-tls-certs\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273399 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8267ea9-4756-45d5-a7bc-8985275cf297-logs\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273455 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgcll\" (UniqueName: \"kubernetes.io/projected/b8267ea9-4756-45d5-a7bc-8985275cf297-kube-api-access-bgcll\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273476 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-combined-ca-bundle\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273576 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-horizon-secret-key\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273598 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8267ea9-4756-45d5-a7bc-8985275cf297-scripts\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273626 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-horizon-tls-certs\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.273650 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8267ea9-4756-45d5-a7bc-8985275cf297-config-data\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.274832 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8267ea9-4756-45d5-a7bc-8985275cf297-config-data\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.275601 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8267ea9-4756-45d5-a7bc-8985275cf297-logs\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.275615 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8267ea9-4756-45d5-a7bc-8985275cf297-scripts\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.281842 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-horizon-secret-key\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.282467 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-horizon-tls-certs\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.282468 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8267ea9-4756-45d5-a7bc-8985275cf297-combined-ca-bundle\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.286613 4710 generic.go:334] "Generic (PLEG): container finished" podID="d9b58541-dcaa-47c4-b256-48ce93a3ab38" containerID="56c067acd68f3ffaa2c02e6108b6721175d00d25a2440ada8b051717cc8a236a" exitCode=0 Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.286705 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cf7wd" event={"ID":"d9b58541-dcaa-47c4-b256-48ce93a3ab38","Type":"ContainerDied","Data":"56c067acd68f3ffaa2c02e6108b6721175d00d25a2440ada8b051717cc8a236a"} Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.305007 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgcll\" (UniqueName: \"kubernetes.io/projected/b8267ea9-4756-45d5-a7bc-8985275cf297-kube-api-access-bgcll\") pod \"horizon-58474b8ddd-nzgx9\" (UID: \"b8267ea9-4756-45d5-a7bc-8985275cf297\") " pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.311521 4710 generic.go:334] "Generic (PLEG): container finished" podID="3d51b0dc-7827-46ff-a400-98758053ef98" containerID="05a23d462401444d029bcda58f52f3b6e480136483d05d729efb4121f5175d26" exitCode=0 Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.311567 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d51b0dc-7827-46ff-a400-98758053ef98","Type":"ContainerDied","Data":"05a23d462401444d029bcda58f52f3b6e480136483d05d729efb4121f5175d26"} Oct 02 13:07:21 crc kubenswrapper[4710]: I1002 13:07:21.312217 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:27 crc kubenswrapper[4710]: I1002 13:07:27.101672 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.151:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.474525 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.475221 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.475449 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59h64h5f5h66bh566h558h88h5fch577h68dh5fch646h58fhc4h67dh668h55ch58bh5d5h5b5h665h56fh679h58fh647hdhcch65h6h5cbh65fh67q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btsqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-65856c9f97-ptvgn_openstack(d887992a-ce0d-4e1c-926d-e1febbde1ca5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.479835 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-65856c9f97-ptvgn" podUID="d887992a-ce0d-4e1c-926d-e1febbde1ca5" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.492475 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.492544 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.492728 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n78hcbh68ch585h5c4h668h68chch67ch97h598hcfh7ch647h658h67chc6h58dh5bbh648hb8h7bh67dh579h9dh675h586hf9h5f9h5f4h657h5b4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99bjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-58dc9c68c-xp6ns_openstack(134d8dfd-37b4-4d67-94d7-aa377e780bae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:07:30 crc kubenswrapper[4710]: E1002 13:07:30.495090 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-58dc9c68c-xp6ns" podUID="134d8dfd-37b4-4d67-94d7-aa377e780bae" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.607075 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.618487 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780511 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d51b0dc-7827-46ff-a400-98758053ef98-logs\") pod \"3d51b0dc-7827-46ff-a400-98758053ef98\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780560 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skgrq\" (UniqueName: \"kubernetes.io/projected/3d51b0dc-7827-46ff-a400-98758053ef98-kube-api-access-skgrq\") pod \"3d51b0dc-7827-46ff-a400-98758053ef98\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780682 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-credential-keys\") pod \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780705 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-config-data\") pod \"3d51b0dc-7827-46ff-a400-98758053ef98\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780760 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-scripts\") pod \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780796 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-custom-prometheus-ca\") pod \"3d51b0dc-7827-46ff-a400-98758053ef98\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780825 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-combined-ca-bundle\") pod \"3d51b0dc-7827-46ff-a400-98758053ef98\" (UID: \"3d51b0dc-7827-46ff-a400-98758053ef98\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780848 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-fernet-keys\") pod \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780903 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82xhw\" (UniqueName: \"kubernetes.io/projected/d9b58541-dcaa-47c4-b256-48ce93a3ab38-kube-api-access-82xhw\") pod \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780933 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-config-data\") pod \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.780961 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-combined-ca-bundle\") pod \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\" (UID: \"d9b58541-dcaa-47c4-b256-48ce93a3ab38\") " Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.781147 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d51b0dc-7827-46ff-a400-98758053ef98-logs" (OuterVolumeSpecName: "logs") pod "3d51b0dc-7827-46ff-a400-98758053ef98" (UID: "3d51b0dc-7827-46ff-a400-98758053ef98"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.781796 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d51b0dc-7827-46ff-a400-98758053ef98-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.787497 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-scripts" (OuterVolumeSpecName: "scripts") pod "d9b58541-dcaa-47c4-b256-48ce93a3ab38" (UID: "d9b58541-dcaa-47c4-b256-48ce93a3ab38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.787789 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d9b58541-dcaa-47c4-b256-48ce93a3ab38" (UID: "d9b58541-dcaa-47c4-b256-48ce93a3ab38"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.788301 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d9b58541-dcaa-47c4-b256-48ce93a3ab38" (UID: "d9b58541-dcaa-47c4-b256-48ce93a3ab38"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.789198 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d51b0dc-7827-46ff-a400-98758053ef98-kube-api-access-skgrq" (OuterVolumeSpecName: "kube-api-access-skgrq") pod "3d51b0dc-7827-46ff-a400-98758053ef98" (UID: "3d51b0dc-7827-46ff-a400-98758053ef98"). InnerVolumeSpecName "kube-api-access-skgrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.791977 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b58541-dcaa-47c4-b256-48ce93a3ab38-kube-api-access-82xhw" (OuterVolumeSpecName: "kube-api-access-82xhw") pod "d9b58541-dcaa-47c4-b256-48ce93a3ab38" (UID: "d9b58541-dcaa-47c4-b256-48ce93a3ab38"). InnerVolumeSpecName "kube-api-access-82xhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.812347 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-config-data" (OuterVolumeSpecName: "config-data") pod "d9b58541-dcaa-47c4-b256-48ce93a3ab38" (UID: "d9b58541-dcaa-47c4-b256-48ce93a3ab38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.821716 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d51b0dc-7827-46ff-a400-98758053ef98" (UID: "3d51b0dc-7827-46ff-a400-98758053ef98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.826451 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "3d51b0dc-7827-46ff-a400-98758053ef98" (UID: "3d51b0dc-7827-46ff-a400-98758053ef98"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.829850 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9b58541-dcaa-47c4-b256-48ce93a3ab38" (UID: "d9b58541-dcaa-47c4-b256-48ce93a3ab38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.840440 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-config-data" (OuterVolumeSpecName: "config-data") pod "3d51b0dc-7827-46ff-a400-98758053ef98" (UID: "3d51b0dc-7827-46ff-a400-98758053ef98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883236 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skgrq\" (UniqueName: \"kubernetes.io/projected/3d51b0dc-7827-46ff-a400-98758053ef98-kube-api-access-skgrq\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883273 4710 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883284 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883294 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883304 4710 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883364 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d51b0dc-7827-46ff-a400-98758053ef98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883376 4710 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883386 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82xhw\" (UniqueName: \"kubernetes.io/projected/d9b58541-dcaa-47c4-b256-48ce93a3ab38-kube-api-access-82xhw\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883409 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:30 crc kubenswrapper[4710]: I1002 13:07:30.883423 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b58541-dcaa-47c4-b256-48ce93a3ab38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:31 crc kubenswrapper[4710]: E1002 13:07:31.019661 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 02 13:07:31 crc kubenswrapper[4710]: E1002 13:07:31.019713 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 02 13:07:31 crc kubenswrapper[4710]: E1002 13:07:31.019909 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.147:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dbh5f8h557hf9h598h5c8h68ch59bh584h687h66fh94h6hf5h674h67dhch548h655h695h85h566h575h576hd5h84h84h585h7bh56ch666hb4q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crqtp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(aed5f89f-6790-4f1f-9ed2-2f075f8d4647): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.403584 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cf7wd" event={"ID":"d9b58541-dcaa-47c4-b256-48ce93a3ab38","Type":"ContainerDied","Data":"74c8fc1c2fd9efd77b0cea27cc77dc5ad5db5a137cb1cada288e59fecc34a778"} Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.403659 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74c8fc1c2fd9efd77b0cea27cc77dc5ad5db5a137cb1cada288e59fecc34a778" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.403603 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cf7wd" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.405964 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.406806 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3d51b0dc-7827-46ff-a400-98758053ef98","Type":"ContainerDied","Data":"1df5a1f636a7627aa4b90bcf0427e762efec93031ab7ab40b01b2e0e729136cc"} Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.406848 4710 scope.go:117] "RemoveContainer" containerID="05a23d462401444d029bcda58f52f3b6e480136483d05d729efb4121f5175d26" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.486449 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.526004 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.536310 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:31 crc kubenswrapper[4710]: E1002 13:07:31.536794 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b58541-dcaa-47c4-b256-48ce93a3ab38" containerName="keystone-bootstrap" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.536808 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b58541-dcaa-47c4-b256-48ce93a3ab38" containerName="keystone-bootstrap" Oct 02 13:07:31 crc kubenswrapper[4710]: E1002 13:07:31.536838 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.536845 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" Oct 02 13:07:31 crc kubenswrapper[4710]: E1002 13:07:31.536865 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api-log" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.536872 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api-log" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.537075 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.537094 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api-log" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.537110 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b58541-dcaa-47c4-b256-48ce93a3ab38" containerName="keystone-bootstrap" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.538133 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.541292 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.545309 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.713281 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b956e3-372d-40f4-9f31-3028269f4175-logs\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.713451 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.713656 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.713711 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-config-data\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.713784 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbc4d\" (UniqueName: \"kubernetes.io/projected/f1b956e3-372d-40f4-9f31-3028269f4175-kube-api-access-xbc4d\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.723018 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-cf7wd"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.733530 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-cf7wd"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.791171 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-987x9"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.794416 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.805708 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p52kv" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.807513 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.808576 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.808705 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.814405 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-987x9"] Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.816233 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.816347 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.816386 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-config-data\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.816426 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbc4d\" (UniqueName: \"kubernetes.io/projected/f1b956e3-372d-40f4-9f31-3028269f4175-kube-api-access-xbc4d\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.816489 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b956e3-372d-40f4-9f31-3028269f4175-logs\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.818328 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b956e3-372d-40f4-9f31-3028269f4175-logs\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.825268 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.833568 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-config-data\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.836219 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.875331 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbc4d\" (UniqueName: \"kubernetes.io/projected/f1b956e3-372d-40f4-9f31-3028269f4175-kube-api-access-xbc4d\") pod \"watcher-api-0\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.876937 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.917692 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-combined-ca-bundle\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.917795 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-config-data\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.917830 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv7sp\" (UniqueName: \"kubernetes.io/projected/54924b84-6352-4ea1-8e55-35ebc23fc0a0-kube-api-access-vv7sp\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.917877 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-scripts\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.917952 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-fernet-keys\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:31 crc kubenswrapper[4710]: I1002 13:07:31.917986 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-credential-keys\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.020135 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-combined-ca-bundle\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.020207 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-config-data\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.020234 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv7sp\" (UniqueName: \"kubernetes.io/projected/54924b84-6352-4ea1-8e55-35ebc23fc0a0-kube-api-access-vv7sp\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.020282 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-scripts\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.020365 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-fernet-keys\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.020409 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-credential-keys\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.024332 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-scripts\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.024901 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-fernet-keys\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.026117 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-combined-ca-bundle\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.036347 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-config-data\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.036538 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv7sp\" (UniqueName: \"kubernetes.io/projected/54924b84-6352-4ea1-8e55-35ebc23fc0a0-kube-api-access-vv7sp\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.047200 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-credential-keys\") pod \"keystone-bootstrap-987x9\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.102931 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.151:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.221889 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.917719 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d51b0dc-7827-46ff-a400-98758053ef98" path="/var/lib/kubelet/pods/3d51b0dc-7827-46ff-a400-98758053ef98/volumes" Oct 02 13:07:32 crc kubenswrapper[4710]: I1002 13:07:32.918734 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b58541-dcaa-47c4-b256-48ce93a3ab38" path="/var/lib/kubelet/pods/d9b58541-dcaa-47c4-b256-48ce93a3ab38/volumes" Oct 02 13:07:42 crc kubenswrapper[4710]: E1002 13:07:42.717445 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 02 13:07:42 crc kubenswrapper[4710]: E1002 13:07:42.718117 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 02 13:07:42 crc kubenswrapper[4710]: E1002 13:07:42.718302 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.147:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pkj2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-nwcvq_openstack(f53ffe3d-d27a-41dd-a98b-4debca318ec4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:07:42 crc kubenswrapper[4710]: E1002 13:07:42.719638 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-nwcvq" podUID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" Oct 02 13:07:42 crc kubenswrapper[4710]: I1002 13:07:42.872823 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:42 crc kubenswrapper[4710]: I1002 13:07:42.886204 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.036977 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-scripts\") pod \"134d8dfd-37b4-4d67-94d7-aa377e780bae\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037042 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btsqs\" (UniqueName: \"kubernetes.io/projected/d887992a-ce0d-4e1c-926d-e1febbde1ca5-kube-api-access-btsqs\") pod \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037090 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d8dfd-37b4-4d67-94d7-aa377e780bae-logs\") pod \"134d8dfd-37b4-4d67-94d7-aa377e780bae\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037133 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-config-data\") pod \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037207 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/134d8dfd-37b4-4d67-94d7-aa377e780bae-horizon-secret-key\") pod \"134d8dfd-37b4-4d67-94d7-aa377e780bae\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037294 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d887992a-ce0d-4e1c-926d-e1febbde1ca5-logs\") pod \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037358 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-config-data\") pod \"134d8dfd-37b4-4d67-94d7-aa377e780bae\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037388 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-scripts\") pod \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037422 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d887992a-ce0d-4e1c-926d-e1febbde1ca5-horizon-secret-key\") pod \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\" (UID: \"d887992a-ce0d-4e1c-926d-e1febbde1ca5\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037486 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bjj\" (UniqueName: \"kubernetes.io/projected/134d8dfd-37b4-4d67-94d7-aa377e780bae-kube-api-access-99bjj\") pod \"134d8dfd-37b4-4d67-94d7-aa377e780bae\" (UID: \"134d8dfd-37b4-4d67-94d7-aa377e780bae\") " Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037524 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-scripts" (OuterVolumeSpecName: "scripts") pod "134d8dfd-37b4-4d67-94d7-aa377e780bae" (UID: "134d8dfd-37b4-4d67-94d7-aa377e780bae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.037777 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134d8dfd-37b4-4d67-94d7-aa377e780bae-logs" (OuterVolumeSpecName: "logs") pod "134d8dfd-37b4-4d67-94d7-aa377e780bae" (UID: "134d8dfd-37b4-4d67-94d7-aa377e780bae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.038021 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/134d8dfd-37b4-4d67-94d7-aa377e780bae-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.038046 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.038076 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d887992a-ce0d-4e1c-926d-e1febbde1ca5-logs" (OuterVolumeSpecName: "logs") pod "d887992a-ce0d-4e1c-926d-e1febbde1ca5" (UID: "d887992a-ce0d-4e1c-926d-e1febbde1ca5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.038280 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-scripts" (OuterVolumeSpecName: "scripts") pod "d887992a-ce0d-4e1c-926d-e1febbde1ca5" (UID: "d887992a-ce0d-4e1c-926d-e1febbde1ca5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.038781 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-config-data" (OuterVolumeSpecName: "config-data") pod "134d8dfd-37b4-4d67-94d7-aa377e780bae" (UID: "134d8dfd-37b4-4d67-94d7-aa377e780bae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.039146 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-config-data" (OuterVolumeSpecName: "config-data") pod "d887992a-ce0d-4e1c-926d-e1febbde1ca5" (UID: "d887992a-ce0d-4e1c-926d-e1febbde1ca5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.041881 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/134d8dfd-37b4-4d67-94d7-aa377e780bae-kube-api-access-99bjj" (OuterVolumeSpecName: "kube-api-access-99bjj") pod "134d8dfd-37b4-4d67-94d7-aa377e780bae" (UID: "134d8dfd-37b4-4d67-94d7-aa377e780bae"). InnerVolumeSpecName "kube-api-access-99bjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.042414 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/134d8dfd-37b4-4d67-94d7-aa377e780bae-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "134d8dfd-37b4-4d67-94d7-aa377e780bae" (UID: "134d8dfd-37b4-4d67-94d7-aa377e780bae"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.042509 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d887992a-ce0d-4e1c-926d-e1febbde1ca5-kube-api-access-btsqs" (OuterVolumeSpecName: "kube-api-access-btsqs") pod "d887992a-ce0d-4e1c-926d-e1febbde1ca5" (UID: "d887992a-ce0d-4e1c-926d-e1febbde1ca5"). InnerVolumeSpecName "kube-api-access-btsqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.062184 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d887992a-ce0d-4e1c-926d-e1febbde1ca5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d887992a-ce0d-4e1c-926d-e1febbde1ca5" (UID: "d887992a-ce0d-4e1c-926d-e1febbde1ca5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139027 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/134d8dfd-37b4-4d67-94d7-aa377e780bae-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139061 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139072 4710 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d887992a-ce0d-4e1c-926d-e1febbde1ca5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139083 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bjj\" (UniqueName: \"kubernetes.io/projected/134d8dfd-37b4-4d67-94d7-aa377e780bae-kube-api-access-99bjj\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139092 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btsqs\" (UniqueName: \"kubernetes.io/projected/d887992a-ce0d-4e1c-926d-e1febbde1ca5-kube-api-access-btsqs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139100 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d887992a-ce0d-4e1c-926d-e1febbde1ca5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139109 4710 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/134d8dfd-37b4-4d67-94d7-aa377e780bae-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.139117 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d887992a-ce0d-4e1c-926d-e1febbde1ca5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.515294 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65856c9f97-ptvgn" event={"ID":"d887992a-ce0d-4e1c-926d-e1febbde1ca5","Type":"ContainerDied","Data":"c99b4a33af205726d0a02c88c3152750e4424deb5864667d42e07e1d1199996b"} Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.515369 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65856c9f97-ptvgn" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.517087 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dc9c68c-xp6ns" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.517144 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dc9c68c-xp6ns" event={"ID":"134d8dfd-37b4-4d67-94d7-aa377e780bae","Type":"ContainerDied","Data":"baec0231e1b89afa894a6a06b2f22dee2bee98108c228e89ce12da464bd8cbc1"} Oct 02 13:07:43 crc kubenswrapper[4710]: E1002 13:07:43.519436 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-nwcvq" podUID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.615971 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58dc9c68c-xp6ns"] Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.631216 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58dc9c68c-xp6ns"] Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.645875 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65856c9f97-ptvgn"] Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.652534 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65856c9f97-ptvgn"] Oct 02 13:07:43 crc kubenswrapper[4710]: I1002 13:07:43.943980 4710 scope.go:117] "RemoveContainer" containerID="5a44308221361954713d3579f7f1efa3d323e878ee244751010bc754ede83449" Oct 02 13:07:43 crc kubenswrapper[4710]: E1002 13:07:43.944303 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 02 13:07:43 crc kubenswrapper[4710]: E1002 13:07:43.944331 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 02 13:07:43 crc kubenswrapper[4710]: E1002 13:07:43.944454 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.147:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hrzkv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-4m9cc_openstack(6152e5af-f395-4bca-9aa1-19d17ec0ffa2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:07:43 crc kubenswrapper[4710]: E1002 13:07:43.945740 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-4m9cc" podUID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.376166 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-588846bfb9-5mhkc"] Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.492138 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58474b8ddd-nzgx9"] Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.502288 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f85867556-bdslh"] Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.560539 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-588846bfb9-5mhkc" event={"ID":"6b39391f-b6e4-4275-941b-63484ea90c69","Type":"ContainerStarted","Data":"3aa5b085b06abaff31c452b9924bbf80c23284695fdd63edd1eca89839e4ddf9"} Oct 02 13:07:44 crc kubenswrapper[4710]: E1002 13:07:44.572260 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-4m9cc" podUID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" Oct 02 13:07:44 crc kubenswrapper[4710]: W1002 13:07:44.586645 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00020afa_1e01_4a2c_924e_2aa55bac0e85.slice/crio-96556850eff82ad42e377905b1f71da4a72c48cac826b297323ecf1b5f654856 WatchSource:0}: Error finding container 96556850eff82ad42e377905b1f71da4a72c48cac826b297323ecf1b5f654856: Status 404 returned error can't find the container with id 96556850eff82ad42e377905b1f71da4a72c48cac826b297323ecf1b5f654856 Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.801480 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-987x9"] Oct 02 13:07:44 crc kubenswrapper[4710]: W1002 13:07:44.814811 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54924b84_6352_4ea1_8e55_35ebc23fc0a0.slice/crio-f416d5d426ad5f05cdb939ac75cee3a87aec0850d8dec052a2c4cc0185ae53fe WatchSource:0}: Error finding container f416d5d426ad5f05cdb939ac75cee3a87aec0850d8dec052a2c4cc0185ae53fe: Status 404 returned error can't find the container with id f416d5d426ad5f05cdb939ac75cee3a87aec0850d8dec052a2c4cc0185ae53fe Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.929328 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="134d8dfd-37b4-4d67-94d7-aa377e780bae" path="/var/lib/kubelet/pods/134d8dfd-37b4-4d67-94d7-aa377e780bae/volumes" Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.929809 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d887992a-ce0d-4e1c-926d-e1febbde1ca5" path="/var/lib/kubelet/pods/d887992a-ce0d-4e1c-926d-e1febbde1ca5/volumes" Oct 02 13:07:44 crc kubenswrapper[4710]: I1002 13:07:44.930168 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.623399 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58474b8ddd-nzgx9" event={"ID":"b8267ea9-4756-45d5-a7bc-8985275cf297","Type":"ContainerStarted","Data":"1208464b7dfd9c12d0b51aa8a6cd65dc43cf329de9a20c3c8e294ac7c6afb278"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.623941 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58474b8ddd-nzgx9" event={"ID":"b8267ea9-4756-45d5-a7bc-8985275cf297","Type":"ContainerStarted","Data":"3c8a5cf9c9dfee3d27c7150a02a3d762b2e630adef68ad4626d36143b776e63d"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.623955 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58474b8ddd-nzgx9" event={"ID":"b8267ea9-4756-45d5-a7bc-8985275cf297","Type":"ContainerStarted","Data":"8248a7d89e46027070db3957314618c96e1a11bdfa05268a013e2660c75c553a"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.634357 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"b868ba48-07b1-4a8e-8a4a-3c129840795c","Type":"ContainerStarted","Data":"6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.638396 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849796779c-f5465" event={"ID":"6d28154a-1fc9-4d45-bb67-ecac912c5d01","Type":"ContainerStarted","Data":"d363b7e81e22cd40e6805e7eb70d6bbee2e985e344a8ad40a9a201cff48832f9"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.639485 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.644377 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9e4fe495-9150-420c-8d44-e56e8808b048","Type":"ContainerStarted","Data":"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.644499 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-httpd" containerID="cri-o://890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3" gracePeriod=30 Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.644475 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-log" containerID="cri-o://91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f" gracePeriod=30 Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.654491 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-987x9" event={"ID":"54924b84-6352-4ea1-8e55-35ebc23fc0a0","Type":"ContainerStarted","Data":"0699a8b45c702e7fd6ac9e39accb03b239a91dbb8189fd8c854e6048a25d80d9"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.654538 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-987x9" event={"ID":"54924b84-6352-4ea1-8e55-35ebc23fc0a0","Type":"ContainerStarted","Data":"f416d5d426ad5f05cdb939ac75cee3a87aec0850d8dec052a2c4cc0185ae53fe"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.687878 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-58474b8ddd-nzgx9" podStartSLOduration=25.569054485 podStartE2EDuration="25.687855115s" podCreationTimestamp="2025-10-02 13:07:20 +0000 UTC" firstStartedPulling="2025-10-02 13:07:44.622226913 +0000 UTC m=+1108.728637796" lastFinishedPulling="2025-10-02 13:07:44.741027543 +0000 UTC m=+1108.847438426" observedRunningTime="2025-10-02 13:07:45.654372522 +0000 UTC m=+1109.760783415" watchObservedRunningTime="2025-10-02 13:07:45.687855115 +0000 UTC m=+1109.794265998" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.694640 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-588846bfb9-5mhkc" event={"ID":"6b39391f-b6e4-4275-941b-63484ea90c69","Type":"ContainerStarted","Data":"2a6b248d51fa9075d87a5aba4358c9b232b87f161558f849071c0ec412f1348a"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.694683 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-588846bfb9-5mhkc" event={"ID":"6b39391f-b6e4-4275-941b-63484ea90c69","Type":"ContainerStarted","Data":"491cc9b96d286224c8c387787f38c76742c3e9412da4f7ad4fec674adbc334a7"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.695223 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-588846bfb9-5mhkc" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon-log" containerID="cri-o://491cc9b96d286224c8c387787f38c76742c3e9412da4f7ad4fec674adbc334a7" gracePeriod=30 Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.695585 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-588846bfb9-5mhkc" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon" containerID="cri-o://2a6b248d51fa9075d87a5aba4358c9b232b87f161558f849071c0ec412f1348a" gracePeriod=30 Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.716970 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=17.111229237 podStartE2EDuration="34.716950937s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="2025-10-02 13:07:13.396579346 +0000 UTC m=+1077.502990229" lastFinishedPulling="2025-10-02 13:07:31.002301046 +0000 UTC m=+1095.108711929" observedRunningTime="2025-10-02 13:07:45.672384146 +0000 UTC m=+1109.778795029" watchObservedRunningTime="2025-10-02 13:07:45.716950937 +0000 UTC m=+1109.823361820" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.735590 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f85867556-bdslh" event={"ID":"00020afa-1e01-4a2c-924e-2aa55bac0e85","Type":"ContainerStarted","Data":"2b4a4bb8893f083849330b9ab1404fe925487d732359bb18512233446d936a8b"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.735649 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f85867556-bdslh" event={"ID":"00020afa-1e01-4a2c-924e-2aa55bac0e85","Type":"ContainerStarted","Data":"a42ca0c2045d5f637fdd39fdecb14ac1b0dd3e6fe65cc97a6a353834d22c6f67"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.735663 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f85867556-bdslh" event={"ID":"00020afa-1e01-4a2c-924e-2aa55bac0e85","Type":"ContainerStarted","Data":"96556850eff82ad42e377905b1f71da4a72c48cac826b297323ecf1b5f654856"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.747001 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dqkk2" event={"ID":"6216b148-2847-437c-8437-1f574973fb87","Type":"ContainerStarted","Data":"85cae4838ac68ae9e6883f3dc5e4fe695b34b8fe3135f482a2d580ac25cec23c"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.747088 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=33.747076966 podStartE2EDuration="33.747076966s" podCreationTimestamp="2025-10-02 13:07:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:45.710089035 +0000 UTC m=+1109.816499918" watchObservedRunningTime="2025-10-02 13:07:45.747076966 +0000 UTC m=+1109.853487849" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.759626 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-849796779c-f5465" podStartSLOduration=34.759609241 podStartE2EDuration="34.759609241s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:45.741135686 +0000 UTC m=+1109.847546569" watchObservedRunningTime="2025-10-02 13:07:45.759609241 +0000 UTC m=+1109.866020124" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.781356 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-588846bfb9-5mhkc" podStartSLOduration=30.534617538 podStartE2EDuration="30.781342368s" podCreationTimestamp="2025-10-02 13:07:15 +0000 UTC" firstStartedPulling="2025-10-02 13:07:44.469655952 +0000 UTC m=+1108.576066835" lastFinishedPulling="2025-10-02 13:07:44.716380782 +0000 UTC m=+1108.822791665" observedRunningTime="2025-10-02 13:07:45.759692983 +0000 UTC m=+1109.866103886" watchObservedRunningTime="2025-10-02 13:07:45.781342368 +0000 UTC m=+1109.887753251" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.782573 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerStarted","Data":"c65aefbb3a54f1ff3f30d8267f4fb52935f8b31fe351a22c4c0e9518a26b77cf"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.807103 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerStarted","Data":"1dc130c3972d51c15bd9ed2746f526598c726a4f4dcc16e5440157a94ad252e1"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.816955 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f85867556-bdslh" podStartSLOduration=25.677221957 podStartE2EDuration="25.816938014s" podCreationTimestamp="2025-10-02 13:07:20 +0000 UTC" firstStartedPulling="2025-10-02 13:07:44.595316285 +0000 UTC m=+1108.701727168" lastFinishedPulling="2025-10-02 13:07:44.735032342 +0000 UTC m=+1108.841443225" observedRunningTime="2025-10-02 13:07:45.807142127 +0000 UTC m=+1109.913553000" watchObservedRunningTime="2025-10-02 13:07:45.816938014 +0000 UTC m=+1109.923348897" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.824081 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a","Type":"ContainerStarted","Data":"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.828669 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f1b956e3-372d-40f4-9f31-3028269f4175","Type":"ContainerStarted","Data":"ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.828717 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f1b956e3-372d-40f4-9f31-3028269f4175","Type":"ContainerStarted","Data":"57c9c8a7f1fe0212344c4b9f1d0883ef461b81f9ba7faf0de1eae3547af61c5e"} Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.829937 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.835996 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.166:9322/\": dial tcp 10.217.0.166:9322: connect: connection refused" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.845231 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-987x9" podStartSLOduration=14.845210866 podStartE2EDuration="14.845210866s" podCreationTimestamp="2025-10-02 13:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:45.840804465 +0000 UTC m=+1109.947215348" watchObservedRunningTime="2025-10-02 13:07:45.845210866 +0000 UTC m=+1109.951621749" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.851090 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.863215 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dqkk2" podStartSLOduration=5.497817784 podStartE2EDuration="34.863197518s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="2025-10-02 13:07:13.351169463 +0000 UTC m=+1077.457580346" lastFinishedPulling="2025-10-02 13:07:42.716549207 +0000 UTC m=+1106.822960080" observedRunningTime="2025-10-02 13:07:45.858214323 +0000 UTC m=+1109.964625206" watchObservedRunningTime="2025-10-02 13:07:45.863197518 +0000 UTC m=+1109.969608401" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.884489 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=17.122708457 podStartE2EDuration="34.884471544s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="2025-10-02 13:07:13.240826256 +0000 UTC m=+1077.347237139" lastFinishedPulling="2025-10-02 13:07:31.002589343 +0000 UTC m=+1095.109000226" observedRunningTime="2025-10-02 13:07:45.881465338 +0000 UTC m=+1109.987876241" watchObservedRunningTime="2025-10-02 13:07:45.884471544 +0000 UTC m=+1109.990882427" Oct 02 13:07:45 crc kubenswrapper[4710]: I1002 13:07:45.916881 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=14.916859409 podStartE2EDuration="14.916859409s" podCreationTimestamp="2025-10-02 13:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:45.913002532 +0000 UTC m=+1110.019413415" watchObservedRunningTime="2025-10-02 13:07:45.916859409 +0000 UTC m=+1110.023270302" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.553436 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567286 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-combined-ca-bundle\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567408 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-logs\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567462 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-config-data\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567571 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-scripts\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567596 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-public-tls-certs\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567629 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-httpd-run\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567651 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.567691 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-254sx\" (UniqueName: \"kubernetes.io/projected/9e4fe495-9150-420c-8d44-e56e8808b048-kube-api-access-254sx\") pod \"9e4fe495-9150-420c-8d44-e56e8808b048\" (UID: \"9e4fe495-9150-420c-8d44-e56e8808b048\") " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.568012 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-logs" (OuterVolumeSpecName: "logs") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.568341 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.568819 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.576012 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e4fe495-9150-420c-8d44-e56e8808b048-kube-api-access-254sx" (OuterVolumeSpecName: "kube-api-access-254sx") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "kube-api-access-254sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.584473 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.587910 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-scripts" (OuterVolumeSpecName: "scripts") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.670401 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.670435 4710 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e4fe495-9150-420c-8d44-e56e8808b048-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.670458 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.670467 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-254sx\" (UniqueName: \"kubernetes.io/projected/9e4fe495-9150-420c-8d44-e56e8808b048-kube-api-access-254sx\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.681977 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-config-data" (OuterVolumeSpecName: "config-data") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.684007 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.718704 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.722046 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9e4fe495-9150-420c-8d44-e56e8808b048" (UID: "9e4fe495-9150-420c-8d44-e56e8808b048"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.774291 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.774502 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.774584 4710 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e4fe495-9150-420c-8d44-e56e8808b048-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.774656 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.844251 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a","Type":"ContainerStarted","Data":"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc"} Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.850629 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f1b956e3-372d-40f4-9f31-3028269f4175","Type":"ContainerStarted","Data":"eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1"} Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.855525 4710 generic.go:334] "Generic (PLEG): container finished" podID="9e4fe495-9150-420c-8d44-e56e8808b048" containerID="890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3" exitCode=0 Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.855557 4710 generic.go:334] "Generic (PLEG): container finished" podID="9e4fe495-9150-420c-8d44-e56e8808b048" containerID="91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f" exitCode=143 Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.856259 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.857706 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9e4fe495-9150-420c-8d44-e56e8808b048","Type":"ContainerDied","Data":"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3"} Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.857750 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9e4fe495-9150-420c-8d44-e56e8808b048","Type":"ContainerDied","Data":"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f"} Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.857764 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9e4fe495-9150-420c-8d44-e56e8808b048","Type":"ContainerDied","Data":"71316dff5f4aa8d4ebbb8ea92b88ad74f1c5892e8fc7fb32bf906cd4408320da"} Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.857797 4710 scope.go:117] "RemoveContainer" containerID="890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.881085 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.937623 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.962315 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.973853 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:46 crc kubenswrapper[4710]: E1002 13:07:46.974461 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-httpd" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.974479 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-httpd" Oct 02 13:07:46 crc kubenswrapper[4710]: E1002 13:07:46.974491 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-log" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.974503 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-log" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.974770 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-httpd" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.974811 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" containerName="glance-log" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.975834 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.983983 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.984039 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 13:07:46 crc kubenswrapper[4710]: I1002 13:07:46.984168 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.082759 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.082850 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.082872 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.082894 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-logs\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.082998 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.083055 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.083089 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66244\" (UniqueName: \"kubernetes.io/projected/cd4716d4-8cdb-4912-9262-caca477c119a-kube-api-access-66244\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.083112 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.137151 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184513 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184578 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184618 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184639 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66244\" (UniqueName: \"kubernetes.io/projected/cd4716d4-8cdb-4912-9262-caca477c119a-kube-api-access-66244\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184703 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184750 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184792 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.184818 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-logs\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.185140 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.187452 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.190089 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.191979 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.192978 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.192988 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-logs\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.209416 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.216196 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66244\" (UniqueName: \"kubernetes.io/projected/cd4716d4-8cdb-4912-9262-caca477c119a-kube-api-access-66244\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.233118 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.301571 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.866515 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-log" containerID="cri-o://d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe" gracePeriod=30 Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.866679 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-httpd" containerID="cri-o://b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc" gracePeriod=30 Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.887935 4710 scope.go:117] "RemoveContainer" containerID="91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f" Oct 02 13:07:47 crc kubenswrapper[4710]: I1002 13:07:47.908114 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=36.908096739 podStartE2EDuration="36.908096739s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:47.90694567 +0000 UTC m=+1112.013356553" watchObservedRunningTime="2025-10-02 13:07:47.908096739 +0000 UTC m=+1112.014507622" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.168977 4710 scope.go:117] "RemoveContainer" containerID="890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3" Oct 02 13:07:48 crc kubenswrapper[4710]: E1002 13:07:48.172254 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3\": container with ID starting with 890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3 not found: ID does not exist" containerID="890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.172335 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3"} err="failed to get container status \"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3\": rpc error: code = NotFound desc = could not find container \"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3\": container with ID starting with 890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3 not found: ID does not exist" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.172376 4710 scope.go:117] "RemoveContainer" containerID="91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f" Oct 02 13:07:48 crc kubenswrapper[4710]: E1002 13:07:48.173040 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f\": container with ID starting with 91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f not found: ID does not exist" containerID="91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.173082 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f"} err="failed to get container status \"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f\": rpc error: code = NotFound desc = could not find container \"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f\": container with ID starting with 91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f not found: ID does not exist" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.173106 4710 scope.go:117] "RemoveContainer" containerID="890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.174060 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3"} err="failed to get container status \"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3\": rpc error: code = NotFound desc = could not find container \"890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3\": container with ID starting with 890e89ca54833a1ee4eee7a1cda8d0ae9afc623e63faf264bdd1213ef66787f3 not found: ID does not exist" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.174102 4710 scope.go:117] "RemoveContainer" containerID="91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.174481 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f"} err="failed to get container status \"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f\": rpc error: code = NotFound desc = could not find container \"91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f\": container with ID starting with 91e9f8be40770f14606a67c0be49325876149dfb87b7b5f847aee6555f479c2f not found: ID does not exist" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.491316 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.758310 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.848903 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-internal-tls-certs\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849071 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849098 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-httpd-run\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849138 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-logs\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849205 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-scripts\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849297 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-config-data\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849350 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-combined-ca-bundle\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.849383 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzjxp\" (UniqueName: \"kubernetes.io/projected/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-kube-api-access-gzjxp\") pod \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\" (UID: \"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a\") " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.850215 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-logs" (OuterVolumeSpecName: "logs") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.850528 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.860660 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-scripts" (OuterVolumeSpecName: "scripts") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.860755 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.874956 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-kube-api-access-gzjxp" (OuterVolumeSpecName: "kube-api-access-gzjxp") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "kube-api-access-gzjxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.901050 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd4716d4-8cdb-4912-9262-caca477c119a","Type":"ContainerStarted","Data":"a3fa776ca6ee57aab4a997b9f3dbf1c18b29c06a2545d570b1f22d3865487c61"} Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.908221 4710 generic.go:334] "Generic (PLEG): container finished" podID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerID="b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc" exitCode=0 Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.908252 4710 generic.go:334] "Generic (PLEG): container finished" podID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerID="d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe" exitCode=143 Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.908326 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.910022 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.920386 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.941019 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e4fe495-9150-420c-8d44-e56e8808b048" path="/var/lib/kubelet/pods/9e4fe495-9150-420c-8d44-e56e8808b048/volumes" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.952550 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.952608 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzjxp\" (UniqueName: \"kubernetes.io/projected/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-kube-api-access-gzjxp\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.952638 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.952651 4710 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.952663 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.952691 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.962737 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-config-data" (OuterVolumeSpecName: "config-data") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.978393 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a","Type":"ContainerDied","Data":"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc"} Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.978458 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a","Type":"ContainerDied","Data":"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe"} Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.978471 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1edb453e-c0a3-4639-96fc-d78a7c9c9e0a","Type":"ContainerDied","Data":"e05ea0d02c95783057d7aa844d6b5bc208cc9799f31e8c1f4dc1f653fcbdf88d"} Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.978514 4710 scope.go:117] "RemoveContainer" containerID="b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc" Oct 02 13:07:48 crc kubenswrapper[4710]: I1002 13:07:48.990065 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.007021 4710 scope.go:117] "RemoveContainer" containerID="d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.016879 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" (UID: "1edb453e-c0a3-4639-96fc-d78a7c9c9e0a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.043168 4710 scope.go:117] "RemoveContainer" containerID="b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc" Oct 02 13:07:49 crc kubenswrapper[4710]: E1002 13:07:49.044836 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc\": container with ID starting with b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc not found: ID does not exist" containerID="b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.044890 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc"} err="failed to get container status \"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc\": rpc error: code = NotFound desc = could not find container \"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc\": container with ID starting with b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc not found: ID does not exist" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.044911 4710 scope.go:117] "RemoveContainer" containerID="d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe" Oct 02 13:07:49 crc kubenswrapper[4710]: E1002 13:07:49.045489 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe\": container with ID starting with d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe not found: ID does not exist" containerID="d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.045517 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe"} err="failed to get container status \"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe\": rpc error: code = NotFound desc = could not find container \"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe\": container with ID starting with d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe not found: ID does not exist" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.045593 4710 scope.go:117] "RemoveContainer" containerID="b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.045958 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc"} err="failed to get container status \"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc\": rpc error: code = NotFound desc = could not find container \"b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc\": container with ID starting with b285862be8f473d64f0104940529cc0167e1347c161ba6c0d36e50eb9f12bcfc not found: ID does not exist" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.045984 4710 scope.go:117] "RemoveContainer" containerID="d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.046193 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe"} err="failed to get container status \"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe\": rpc error: code = NotFound desc = could not find container \"d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe\": container with ID starting with d208ddad5e23354944fea7fe25452444bebe20455d3822a6cef4578241c64afe not found: ID does not exist" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.055092 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.055137 4710 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.055153 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.257917 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.270125 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.284003 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:49 crc kubenswrapper[4710]: E1002 13:07:49.284435 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-log" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.284451 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-log" Oct 02 13:07:49 crc kubenswrapper[4710]: E1002 13:07:49.284479 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-httpd" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.284488 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-httpd" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.284704 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-httpd" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.284727 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" containerName="glance-log" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.285950 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.289638 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.290078 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.298694 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.464481 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.464561 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.464667 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-logs\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.466879 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.466993 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbk4v\" (UniqueName: \"kubernetes.io/projected/48daf522-43e6-4fa5-a526-ee522fda7ae2-kube-api-access-mbk4v\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.467060 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.467080 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.467280 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568459 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-logs\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568502 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568551 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbk4v\" (UniqueName: \"kubernetes.io/projected/48daf522-43e6-4fa5-a526-ee522fda7ae2-kube-api-access-mbk4v\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568585 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568610 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568647 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568675 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.568700 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.569173 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-logs\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.569611 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.569665 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.575466 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.575716 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.576370 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.587973 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbk4v\" (UniqueName: \"kubernetes.io/projected/48daf522-43e6-4fa5-a526-ee522fda7ae2-kube-api-access-mbk4v\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.599012 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.611011 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.647246 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:49 crc kubenswrapper[4710]: I1002 13:07:49.922310 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd4716d4-8cdb-4912-9262-caca477c119a","Type":"ContainerStarted","Data":"bf7c7ff0c599541828e59be2370327c35086721fb238d5b356aeef192b120573"} Oct 02 13:07:50 crc kubenswrapper[4710]: I1002 13:07:50.457334 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 13:07:50 crc kubenswrapper[4710]: I1002 13:07:50.920379 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1edb453e-c0a3-4639-96fc-d78a7c9c9e0a" path="/var/lib/kubelet/pods/1edb453e-c0a3-4639-96fc-d78a7c9c9e0a/volumes" Oct 02 13:07:50 crc kubenswrapper[4710]: I1002 13:07:50.936618 4710 generic.go:334] "Generic (PLEG): container finished" podID="54924b84-6352-4ea1-8e55-35ebc23fc0a0" containerID="0699a8b45c702e7fd6ac9e39accb03b239a91dbb8189fd8c854e6048a25d80d9" exitCode=0 Oct 02 13:07:50 crc kubenswrapper[4710]: I1002 13:07:50.936691 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-987x9" event={"ID":"54924b84-6352-4ea1-8e55-35ebc23fc0a0","Type":"ContainerDied","Data":"0699a8b45c702e7fd6ac9e39accb03b239a91dbb8189fd8c854e6048a25d80d9"} Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.135660 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.135725 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.313532 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.313941 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.878034 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.892293 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.950118 4710 generic.go:334] "Generic (PLEG): container finished" podID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerID="c65aefbb3a54f1ff3f30d8267f4fb52935f8b31fe351a22c4c0e9518a26b77cf" exitCode=1 Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.950449 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerDied","Data":"c65aefbb3a54f1ff3f30d8267f4fb52935f8b31fe351a22c4c0e9518a26b77cf"} Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.952393 4710 scope.go:117] "RemoveContainer" containerID="c65aefbb3a54f1ff3f30d8267f4fb52935f8b31fe351a22c4c0e9518a26b77cf" Oct 02 13:07:51 crc kubenswrapper[4710]: I1002 13:07:51.957277 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.092355 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.092936 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.137132 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.191113 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.410933 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.503531 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bb78f7fbf-gkh8q"] Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.503879 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerName="dnsmasq-dns" containerID="cri-o://d069eaf80733544e7afcb6b1d6ccd56e4a4d04b44060b04ae2c7e9d2bd1c44bf" gracePeriod=10 Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.961621 4710 generic.go:334] "Generic (PLEG): container finished" podID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerID="d069eaf80733544e7afcb6b1d6ccd56e4a4d04b44060b04ae2c7e9d2bd1c44bf" exitCode=0 Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.961705 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" event={"ID":"6af2e16f-eced-4e5f-90b8-e1fffd0d6573","Type":"ContainerDied","Data":"d069eaf80733544e7afcb6b1d6ccd56e4a4d04b44060b04ae2c7e9d2bd1c44bf"} Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.965275 4710 generic.go:334] "Generic (PLEG): container finished" podID="6216b148-2847-437c-8437-1f574973fb87" containerID="85cae4838ac68ae9e6883f3dc5e4fe695b34b8fe3135f482a2d580ac25cec23c" exitCode=0 Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.965353 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dqkk2" event={"ID":"6216b148-2847-437c-8437-1f574973fb87","Type":"ContainerDied","Data":"85cae4838ac68ae9e6883f3dc5e4fe695b34b8fe3135f482a2d580ac25cec23c"} Oct 02 13:07:52 crc kubenswrapper[4710]: I1002 13:07:52.996050 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 02 13:07:53 crc kubenswrapper[4710]: I1002 13:07:53.033802 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.697645 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.703389 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882180 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-combined-ca-bundle\") pod \"6216b148-2847-437c-8437-1f574973fb87\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882599 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-credential-keys\") pod \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882654 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-scripts\") pod \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882782 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-combined-ca-bundle\") pod \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882809 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-config-data\") pod \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882864 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kssg4\" (UniqueName: \"kubernetes.io/projected/6216b148-2847-437c-8437-1f574973fb87-kube-api-access-kssg4\") pod \"6216b148-2847-437c-8437-1f574973fb87\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882904 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6216b148-2847-437c-8437-1f574973fb87-logs\") pod \"6216b148-2847-437c-8437-1f574973fb87\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882956 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-scripts\") pod \"6216b148-2847-437c-8437-1f574973fb87\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.882995 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv7sp\" (UniqueName: \"kubernetes.io/projected/54924b84-6352-4ea1-8e55-35ebc23fc0a0-kube-api-access-vv7sp\") pod \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.883019 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-fernet-keys\") pod \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\" (UID: \"54924b84-6352-4ea1-8e55-35ebc23fc0a0\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.883043 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-config-data\") pod \"6216b148-2847-437c-8437-1f574973fb87\" (UID: \"6216b148-2847-437c-8437-1f574973fb87\") " Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.890377 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6216b148-2847-437c-8437-1f574973fb87-logs" (OuterVolumeSpecName: "logs") pod "6216b148-2847-437c-8437-1f574973fb87" (UID: "6216b148-2847-437c-8437-1f574973fb87"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.921319 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6216b148-2847-437c-8437-1f574973fb87-kube-api-access-kssg4" (OuterVolumeSpecName: "kube-api-access-kssg4") pod "6216b148-2847-437c-8437-1f574973fb87" (UID: "6216b148-2847-437c-8437-1f574973fb87"). InnerVolumeSpecName "kube-api-access-kssg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.921701 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54924b84-6352-4ea1-8e55-35ebc23fc0a0-kube-api-access-vv7sp" (OuterVolumeSpecName: "kube-api-access-vv7sp") pod "54924b84-6352-4ea1-8e55-35ebc23fc0a0" (UID: "54924b84-6352-4ea1-8e55-35ebc23fc0a0"). InnerVolumeSpecName "kube-api-access-vv7sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.927123 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-scripts" (OuterVolumeSpecName: "scripts") pod "6216b148-2847-437c-8437-1f574973fb87" (UID: "6216b148-2847-437c-8437-1f574973fb87"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.937514 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-scripts" (OuterVolumeSpecName: "scripts") pod "54924b84-6352-4ea1-8e55-35ebc23fc0a0" (UID: "54924b84-6352-4ea1-8e55-35ebc23fc0a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.956413 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "54924b84-6352-4ea1-8e55-35ebc23fc0a0" (UID: "54924b84-6352-4ea1-8e55-35ebc23fc0a0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.970468 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "54924b84-6352-4ea1-8e55-35ebc23fc0a0" (UID: "54924b84-6352-4ea1-8e55-35ebc23fc0a0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986086 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kssg4\" (UniqueName: \"kubernetes.io/projected/6216b148-2847-437c-8437-1f574973fb87-kube-api-access-kssg4\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986125 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6216b148-2847-437c-8437-1f574973fb87-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986139 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986150 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv7sp\" (UniqueName: \"kubernetes.io/projected/54924b84-6352-4ea1-8e55-35ebc23fc0a0-kube-api-access-vv7sp\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986161 4710 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986173 4710 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:54 crc kubenswrapper[4710]: I1002 13:07:54.986184 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.019975 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dqkk2" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.020169 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dqkk2" event={"ID":"6216b148-2847-437c-8437-1f574973fb87","Type":"ContainerDied","Data":"9020a48a490dd2d83a5762bc2434ba1b9b3e521cdbae6c9dae39aa7da8719908"} Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.020314 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9020a48a490dd2d83a5762bc2434ba1b9b3e521cdbae6c9dae39aa7da8719908" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.050612 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="b868ba48-07b1-4a8e-8a4a-3c129840795c" containerName="watcher-applier" containerID="cri-o://6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703" gracePeriod=30 Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.050840 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-987x9" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.051260 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-987x9" event={"ID":"54924b84-6352-4ea1-8e55-35ebc23fc0a0","Type":"ContainerDied","Data":"f416d5d426ad5f05cdb939ac75cee3a87aec0850d8dec052a2c4cc0185ae53fe"} Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.051291 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f416d5d426ad5f05cdb939ac75cee3a87aec0850d8dec052a2c4cc0185ae53fe" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.070206 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.098707 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6216b148-2847-437c-8437-1f574973fb87" (UID: "6216b148-2847-437c-8437-1f574973fb87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.115473 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-config-data" (OuterVolumeSpecName: "config-data") pod "54924b84-6352-4ea1-8e55-35ebc23fc0a0" (UID: "54924b84-6352-4ea1-8e55-35ebc23fc0a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.122238 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-config-data" (OuterVolumeSpecName: "config-data") pod "6216b148-2847-437c-8437-1f574973fb87" (UID: "6216b148-2847-437c-8437-1f574973fb87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.139922 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8cd8bc89b-fpjvs"] Oct 02 13:07:55 crc kubenswrapper[4710]: E1002 13:07:55.140422 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerName="init" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140447 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerName="init" Oct 02 13:07:55 crc kubenswrapper[4710]: E1002 13:07:55.140456 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerName="dnsmasq-dns" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140464 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerName="dnsmasq-dns" Oct 02 13:07:55 crc kubenswrapper[4710]: E1002 13:07:55.140489 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54924b84-6352-4ea1-8e55-35ebc23fc0a0" containerName="keystone-bootstrap" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140499 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="54924b84-6352-4ea1-8e55-35ebc23fc0a0" containerName="keystone-bootstrap" Oct 02 13:07:55 crc kubenswrapper[4710]: E1002 13:07:55.140535 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6216b148-2847-437c-8437-1f574973fb87" containerName="placement-db-sync" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140543 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6216b148-2847-437c-8437-1f574973fb87" containerName="placement-db-sync" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140784 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="54924b84-6352-4ea1-8e55-35ebc23fc0a0" containerName="keystone-bootstrap" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140803 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6216b148-2847-437c-8437-1f574973fb87" containerName="placement-db-sync" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.140817 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" containerName="dnsmasq-dns" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.141935 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.147738 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.148089 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.158623 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54924b84-6352-4ea1-8e55-35ebc23fc0a0" (UID: "54924b84-6352-4ea1-8e55-35ebc23fc0a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.197366 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfgkv\" (UniqueName: \"kubernetes.io/projected/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-kube-api-access-pfgkv\") pod \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.197504 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-sb\") pod \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.197571 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-config\") pod \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.197671 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-nb\") pod \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.197704 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-dns-svc\") pod \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\" (UID: \"6af2e16f-eced-4e5f-90b8-e1fffd0d6573\") " Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.198272 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.198294 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54924b84-6352-4ea1-8e55-35ebc23fc0a0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.198305 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.198317 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6216b148-2847-437c-8437-1f574973fb87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.217041 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8cd8bc89b-fpjvs"] Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.224999 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-kube-api-access-pfgkv" (OuterVolumeSpecName: "kube-api-access-pfgkv") pod "6af2e16f-eced-4e5f-90b8-e1fffd0d6573" (UID: "6af2e16f-eced-4e5f-90b8-e1fffd0d6573"). InnerVolumeSpecName "kube-api-access-pfgkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.271260 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-config" (OuterVolumeSpecName: "config") pod "6af2e16f-eced-4e5f-90b8-e1fffd0d6573" (UID: "6af2e16f-eced-4e5f-90b8-e1fffd0d6573"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.299297 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6af2e16f-eced-4e5f-90b8-e1fffd0d6573" (UID: "6af2e16f-eced-4e5f-90b8-e1fffd0d6573"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307177 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-internal-tls-certs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307257 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-config-data\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307383 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-combined-ca-bundle\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307414 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlw7q\" (UniqueName: \"kubernetes.io/projected/ed865d67-a052-4171-8d50-7cd44ab89f7d-kube-api-access-mlw7q\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307481 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-public-tls-certs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307547 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-scripts\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307590 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed865d67-a052-4171-8d50-7cd44ab89f7d-logs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307810 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfgkv\" (UniqueName: \"kubernetes.io/projected/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-kube-api-access-pfgkv\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307836 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.307849 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.315830 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.316250 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6af2e16f-eced-4e5f-90b8-e1fffd0d6573" (UID: "6af2e16f-eced-4e5f-90b8-e1fffd0d6573"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.325555 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6af2e16f-eced-4e5f-90b8-e1fffd0d6573" (UID: "6af2e16f-eced-4e5f-90b8-e1fffd0d6573"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.409933 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-public-tls-certs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410319 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-scripts\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410359 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed865d67-a052-4171-8d50-7cd44ab89f7d-logs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410469 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-internal-tls-certs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410508 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-config-data\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410587 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-combined-ca-bundle\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410624 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlw7q\" (UniqueName: \"kubernetes.io/projected/ed865d67-a052-4171-8d50-7cd44ab89f7d-kube-api-access-mlw7q\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410202 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410952 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api-log" containerID="cri-o://ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e" gracePeriod=30 Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.410686 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.411204 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6af2e16f-eced-4e5f-90b8-e1fffd0d6573-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.411430 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api" containerID="cri-o://eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1" gracePeriod=30 Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.411618 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed865d67-a052-4171-8d50-7cd44ab89f7d-logs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.414849 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-public-tls-certs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.416229 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-internal-tls-certs\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.416786 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-scripts\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.424202 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-combined-ca-bundle\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.431405 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed865d67-a052-4171-8d50-7cd44ab89f7d-config-data\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.441370 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlw7q\" (UniqueName: \"kubernetes.io/projected/ed865d67-a052-4171-8d50-7cd44ab89f7d-kube-api-access-mlw7q\") pod \"placement-8cd8bc89b-fpjvs\" (UID: \"ed865d67-a052-4171-8d50-7cd44ab89f7d\") " pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.476266 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.855212 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7fc95bbdbd-gxqdq"] Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.856999 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.863815 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.864079 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.864251 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.864940 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.865401 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p52kv" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.865545 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 13:07:55 crc kubenswrapper[4710]: I1002 13:07:55.872882 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fc95bbdbd-gxqdq"] Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046313 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-combined-ca-bundle\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046642 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-public-tls-certs\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046702 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-scripts\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046727 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-internal-tls-certs\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046788 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-credential-keys\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046811 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-fernet-keys\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.046841 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-config-data\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.047007 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rrgb\" (UniqueName: \"kubernetes.io/projected/269c359c-bcb5-4a9c-97f2-b66a8ee22449-kube-api-access-2rrgb\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.106775 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" event={"ID":"6af2e16f-eced-4e5f-90b8-e1fffd0d6573","Type":"ContainerDied","Data":"f97ec75f287de9875b9885fa95652d7e83a483817c39d637ee0d56673a3ed9bd"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.106833 4710 scope.go:117] "RemoveContainer" containerID="d069eaf80733544e7afcb6b1d6ccd56e4a4d04b44060b04ae2c7e9d2bd1c44bf" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.106998 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb78f7fbf-gkh8q" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.158035 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8cd8bc89b-fpjvs"] Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161308 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-scripts\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161359 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-internal-tls-certs\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161395 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-credential-keys\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161418 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-fernet-keys\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161447 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-config-data\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161516 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rrgb\" (UniqueName: \"kubernetes.io/projected/269c359c-bcb5-4a9c-97f2-b66a8ee22449-kube-api-access-2rrgb\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161611 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-combined-ca-bundle\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.161648 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-public-tls-certs\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.162348 4710 generic.go:334] "Generic (PLEG): container finished" podID="f1b956e3-372d-40f4-9f31-3028269f4175" containerID="ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e" exitCode=143 Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.162425 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f1b956e3-372d-40f4-9f31-3028269f4175","Type":"ContainerDied","Data":"ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.169157 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-config-data\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.174107 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-fernet-keys\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.176299 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-credential-keys\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.176980 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-internal-tls-certs\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.179798 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-scripts\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.180043 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-public-tls-certs\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.188547 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269c359c-bcb5-4a9c-97f2-b66a8ee22449-combined-ca-bundle\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.189034 4710 generic.go:334] "Generic (PLEG): container finished" podID="978ba8dd-e49f-40cc-a4f9-4dd5658e4001" containerID="258fdf3876f9e30b2aeb09f882a559bc54129f5529eed0bc6e82425ed444a55a" exitCode=0 Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.189143 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-swkp8" event={"ID":"978ba8dd-e49f-40cc-a4f9-4dd5658e4001","Type":"ContainerDied","Data":"258fdf3876f9e30b2aeb09f882a559bc54129f5529eed0bc6e82425ed444a55a"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.212405 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rrgb\" (UniqueName: \"kubernetes.io/projected/269c359c-bcb5-4a9c-97f2-b66a8ee22449-kube-api-access-2rrgb\") pod \"keystone-7fc95bbdbd-gxqdq\" (UID: \"269c359c-bcb5-4a9c-97f2-b66a8ee22449\") " pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.226245 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd4716d4-8cdb-4912-9262-caca477c119a","Type":"ContainerStarted","Data":"c793e42859c6f8f5a3a59a1f2c542ff0e34fcd835bde0fd3100307777c56574e"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.250371 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48daf522-43e6-4fa5-a526-ee522fda7ae2","Type":"ContainerStarted","Data":"d76fb019c7a67cdccccdc7bd92f172efd406ed630190616d3565233ead6366cf"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.256331 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.256313695 podStartE2EDuration="10.256313695s" podCreationTimestamp="2025-10-02 13:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:56.252177191 +0000 UTC m=+1120.358588074" watchObservedRunningTime="2025-10-02 13:07:56.256313695 +0000 UTC m=+1120.362724578" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.278463 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerStarted","Data":"12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.278530 4710 scope.go:117] "RemoveContainer" containerID="dd23a219ffc22b40a4bd6c12295fc46869251d7471604e7b434cec63651b5536" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.299040 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerStarted","Data":"f31d2c3b09a941fffe2619fe6a392f8621451418117fb3f7673b22acebccc053"} Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.325600 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bb78f7fbf-gkh8q"] Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.359186 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bb78f7fbf-gkh8q"] Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.493671 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:56 crc kubenswrapper[4710]: I1002 13:07:56.924304 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6af2e16f-eced-4e5f-90b8-e1fffd0d6573" path="/var/lib/kubelet/pods/6af2e16f-eced-4e5f-90b8-e1fffd0d6573/volumes" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.048243 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fc95bbdbd-gxqdq"] Oct 02 13:07:57 crc kubenswrapper[4710]: W1002 13:07:57.099884 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod269c359c_bcb5_4a9c_97f2_b66a8ee22449.slice/crio-8ca540e504ecb9639e656ef23e105c4352129d008d610257d8901c81a45ffb4a WatchSource:0}: Error finding container 8ca540e504ecb9639e656ef23e105c4352129d008d610257d8901c81a45ffb4a: Status 404 returned error can't find the container with id 8ca540e504ecb9639e656ef23e105c4352129d008d610257d8901c81a45ffb4a Oct 02 13:07:57 crc kubenswrapper[4710]: E1002 13:07:57.168165 4710 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Oct 02 13:07:57 crc kubenswrapper[4710]: E1002 13:07:57.188788 4710 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Oct 02 13:07:57 crc kubenswrapper[4710]: E1002 13:07:57.288636 4710 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Oct 02 13:07:57 crc kubenswrapper[4710]: E1002 13:07:57.288707 4710 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="b868ba48-07b1-4a8e-8a4a-3c129840795c" containerName="watcher-applier" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.304024 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.304078 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.350002 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9322/\": read tcp 10.217.0.2:41720->10.217.0.166:9322: read: connection reset by peer" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.350392 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.166:9322/\": read tcp 10.217.0.2:41714->10.217.0.166:9322: read: connection reset by peer" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.408942 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.410460 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48daf522-43e6-4fa5-a526-ee522fda7ae2","Type":"ContainerStarted","Data":"571b65a30e776db3a5c455b8f64efd6f218c1c142dad47158a753fae976a139c"} Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.418419 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8cd8bc89b-fpjvs" event={"ID":"ed865d67-a052-4171-8d50-7cd44ab89f7d","Type":"ContainerStarted","Data":"3384e4c9533c034685a26e5b86562e97dadefea07fcf95774180ac92009f40bf"} Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.418460 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8cd8bc89b-fpjvs" event={"ID":"ed865d67-a052-4171-8d50-7cd44ab89f7d","Type":"ContainerStarted","Data":"4606276c12d8ec8f19db5b8ff6e38613e989c8f0cf84f4d34e233118ae3f812f"} Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.418469 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8cd8bc89b-fpjvs" event={"ID":"ed865d67-a052-4171-8d50-7cd44ab89f7d","Type":"ContainerStarted","Data":"2d5a189a77cb82a39371d11a3249c0b4a4bb8980cf6835ee481615d671f821b1"} Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.419905 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.420352 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.428358 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fc95bbdbd-gxqdq" event={"ID":"269c359c-bcb5-4a9c-97f2-b66a8ee22449","Type":"ContainerStarted","Data":"8ca540e504ecb9639e656ef23e105c4352129d008d610257d8901c81a45ffb4a"} Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.428398 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.453034 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.460585 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8cd8bc89b-fpjvs" podStartSLOduration=2.4605665070000002 podStartE2EDuration="2.460566507s" podCreationTimestamp="2025-10-02 13:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:57.459255324 +0000 UTC m=+1121.565666217" watchObservedRunningTime="2025-10-02 13:07:57.460566507 +0000 UTC m=+1121.566977390" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.925552 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.946293 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq4q5\" (UniqueName: \"kubernetes.io/projected/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-kube-api-access-fq4q5\") pod \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.946384 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-config\") pod \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.946636 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-combined-ca-bundle\") pod \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\" (UID: \"978ba8dd-e49f-40cc-a4f9-4dd5658e4001\") " Oct 02 13:07:57 crc kubenswrapper[4710]: I1002 13:07:57.956116 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-kube-api-access-fq4q5" (OuterVolumeSpecName: "kube-api-access-fq4q5") pod "978ba8dd-e49f-40cc-a4f9-4dd5658e4001" (UID: "978ba8dd-e49f-40cc-a4f9-4dd5658e4001"). InnerVolumeSpecName "kube-api-access-fq4q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.048023 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.050351 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq4q5\" (UniqueName: \"kubernetes.io/projected/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-kube-api-access-fq4q5\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.067667 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "978ba8dd-e49f-40cc-a4f9-4dd5658e4001" (UID: "978ba8dd-e49f-40cc-a4f9-4dd5658e4001"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.097651 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-config" (OuterVolumeSpecName: "config") pod "978ba8dd-e49f-40cc-a4f9-4dd5658e4001" (UID: "978ba8dd-e49f-40cc-a4f9-4dd5658e4001"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.151835 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-config-data\") pod \"f1b956e3-372d-40f4-9f31-3028269f4175\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.152009 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-combined-ca-bundle\") pod \"f1b956e3-372d-40f4-9f31-3028269f4175\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.152091 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b956e3-372d-40f4-9f31-3028269f4175-logs\") pod \"f1b956e3-372d-40f4-9f31-3028269f4175\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.152213 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-custom-prometheus-ca\") pod \"f1b956e3-372d-40f4-9f31-3028269f4175\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.152264 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbc4d\" (UniqueName: \"kubernetes.io/projected/f1b956e3-372d-40f4-9f31-3028269f4175-kube-api-access-xbc4d\") pod \"f1b956e3-372d-40f4-9f31-3028269f4175\" (UID: \"f1b956e3-372d-40f4-9f31-3028269f4175\") " Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.153856 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.153886 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/978ba8dd-e49f-40cc-a4f9-4dd5658e4001-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.154461 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1b956e3-372d-40f4-9f31-3028269f4175-logs" (OuterVolumeSpecName: "logs") pod "f1b956e3-372d-40f4-9f31-3028269f4175" (UID: "f1b956e3-372d-40f4-9f31-3028269f4175"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.173970 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1b956e3-372d-40f4-9f31-3028269f4175-kube-api-access-xbc4d" (OuterVolumeSpecName: "kube-api-access-xbc4d") pod "f1b956e3-372d-40f4-9f31-3028269f4175" (UID: "f1b956e3-372d-40f4-9f31-3028269f4175"). InnerVolumeSpecName "kube-api-access-xbc4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.244129 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1b956e3-372d-40f4-9f31-3028269f4175" (UID: "f1b956e3-372d-40f4-9f31-3028269f4175"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.252225 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "f1b956e3-372d-40f4-9f31-3028269f4175" (UID: "f1b956e3-372d-40f4-9f31-3028269f4175"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.255457 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.255489 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b956e3-372d-40f4-9f31-3028269f4175-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.255518 4710 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.255529 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbc4d\" (UniqueName: \"kubernetes.io/projected/f1b956e3-372d-40f4-9f31-3028269f4175-kube-api-access-xbc4d\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.343418 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-config-data" (OuterVolumeSpecName: "config-data") pod "f1b956e3-372d-40f4-9f31-3028269f4175" (UID: "f1b956e3-372d-40f4-9f31-3028269f4175"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.357936 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b956e3-372d-40f4-9f31-3028269f4175-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.477585 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-swkp8" event={"ID":"978ba8dd-e49f-40cc-a4f9-4dd5658e4001","Type":"ContainerDied","Data":"3853388b578afd4b9961717af12dc32334e02a6f0f5912b901769ca93fa8a0c5"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.477628 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3853388b578afd4b9961717af12dc32334e02a6f0f5912b901769ca93fa8a0c5" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.477698 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-swkp8" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.478524 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59b4d46b6c-l4bbj"] Oct 02 13:07:58 crc kubenswrapper[4710]: E1002 13:07:58.478906 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api-log" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.478923 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api-log" Oct 02 13:07:58 crc kubenswrapper[4710]: E1002 13:07:58.478949 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978ba8dd-e49f-40cc-a4f9-4dd5658e4001" containerName="neutron-db-sync" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.478957 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="978ba8dd-e49f-40cc-a4f9-4dd5658e4001" containerName="neutron-db-sync" Oct 02 13:07:58 crc kubenswrapper[4710]: E1002 13:07:58.478980 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.478986 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.479150 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api-log" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.479162 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="978ba8dd-e49f-40cc-a4f9-4dd5658e4001" containerName="neutron-db-sync" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.479182 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" containerName="watcher-api" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.521794 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b4d46b6c-l4bbj"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.521895 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.538633 4710 generic.go:334] "Generic (PLEG): container finished" podID="b868ba48-07b1-4a8e-8a4a-3c129840795c" containerID="6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703" exitCode=0 Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.538735 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"b868ba48-07b1-4a8e-8a4a-3c129840795c","Type":"ContainerDied","Data":"6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.551921 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fc95bbdbd-gxqdq" event={"ID":"269c359c-bcb5-4a9c-97f2-b66a8ee22449","Type":"ContainerStarted","Data":"3ef3e7d6b6e123ac2a558d21f91565336c14493893f58456333d8206a6cf07ad"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.552098 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.574414 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nwcvq" event={"ID":"f53ffe3d-d27a-41dd-a98b-4debca318ec4","Type":"ContainerStarted","Data":"712b61b517d05149b02904aa2054f336915e1e5e34db2dae2941851eac765724"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.579348 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-svc\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.579480 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-config\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.579509 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-sb\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.579607 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-nb\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.579667 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhhmn\" (UniqueName: \"kubernetes.io/projected/48923995-d4ba-4562-8db2-26d5d34a0be5-kube-api-access-rhhmn\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.579821 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-swift-storage-0\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.600572 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7fc95bbdbd-gxqdq" podStartSLOduration=3.600549891 podStartE2EDuration="3.600549891s" podCreationTimestamp="2025-10-02 13:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:58.575540561 +0000 UTC m=+1122.681951444" watchObservedRunningTime="2025-10-02 13:07:58.600549891 +0000 UTC m=+1122.706960774" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.621068 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48daf522-43e6-4fa5-a526-ee522fda7ae2","Type":"ContainerStarted","Data":"c3e02d5842f0eff8776966942e67439672947e4cbb6d9da165b5d2a389112b31"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.632420 4710 generic.go:334] "Generic (PLEG): container finished" podID="f1b956e3-372d-40f4-9f31-3028269f4175" containerID="eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1" exitCode=0 Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.633406 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.634916 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f1b956e3-372d-40f4-9f31-3028269f4175","Type":"ContainerDied","Data":"eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.634994 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f1b956e3-372d-40f4-9f31-3028269f4175","Type":"ContainerDied","Data":"57c9c8a7f1fe0212344c4b9f1d0883ef461b81f9ba7faf0de1eae3547af61c5e"} Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.635013 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.635035 4710 scope.go:117] "RemoveContainer" containerID="eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.641549 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-nwcvq" podStartSLOduration=4.239867903 podStartE2EDuration="47.641527902s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="2025-10-02 13:07:13.647048421 +0000 UTC m=+1077.753459304" lastFinishedPulling="2025-10-02 13:07:57.04870842 +0000 UTC m=+1121.155119303" observedRunningTime="2025-10-02 13:07:58.621305053 +0000 UTC m=+1122.727715946" watchObservedRunningTime="2025-10-02 13:07:58.641527902 +0000 UTC m=+1122.747938785" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.681839 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-svc\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.681934 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-sb\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.682229 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-config\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.682267 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-nb\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.682292 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhhmn\" (UniqueName: \"kubernetes.io/projected/48923995-d4ba-4562-8db2-26d5d34a0be5-kube-api-access-rhhmn\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.682350 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-swift-storage-0\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.683576 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-svc\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.684528 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-sb\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.685966 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.685924729 podStartE2EDuration="9.685924729s" podCreationTimestamp="2025-10-02 13:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:07:58.678531473 +0000 UTC m=+1122.784942356" watchObservedRunningTime="2025-10-02 13:07:58.685924729 +0000 UTC m=+1122.792335612" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.688706 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-nb\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.689283 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-swift-storage-0\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.708252 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-config\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.739444 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhhmn\" (UniqueName: \"kubernetes.io/projected/48923995-d4ba-4562-8db2-26d5d34a0be5-kube-api-access-rhhmn\") pod \"dnsmasq-dns-59b4d46b6c-l4bbj\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.747079 4710 scope.go:117] "RemoveContainer" containerID="ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.753848 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f5fcfd868-8dmb9"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.755372 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.765501 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.765734 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.765886 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.766032 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-458k4" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.815930 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f5fcfd868-8dmb9"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.841947 4710 scope.go:117] "RemoveContainer" containerID="eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1" Oct 02 13:07:58 crc kubenswrapper[4710]: E1002 13:07:58.844255 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1\": container with ID starting with eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1 not found: ID does not exist" containerID="eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.844372 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1"} err="failed to get container status \"eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1\": rpc error: code = NotFound desc = could not find container \"eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1\": container with ID starting with eddbdb3be13a07c837d2d1a6c4a4ac1bb18226ee431d8f96ff5f09ecbed23ed1 not found: ID does not exist" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.844478 4710 scope.go:117] "RemoveContainer" containerID="ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e" Oct 02 13:07:58 crc kubenswrapper[4710]: E1002 13:07:58.845884 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e\": container with ID starting with ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e not found: ID does not exist" containerID="ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.845979 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e"} err="failed to get container status \"ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e\": rpc error: code = NotFound desc = could not find container \"ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e\": container with ID starting with ce02dd6adc8d24e094c61d9828fac06ff971876ec55a828b377f6ebb1555d29e not found: ID does not exist" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.872125 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.878298 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.881294 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.887905 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.890728 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.893525 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vwnh\" (UniqueName: \"kubernetes.io/projected/e2022f45-9ec9-483a-bce6-85945738da7f-kube-api-access-4vwnh\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.893656 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-httpd-config\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.893697 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-ovndb-tls-certs\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.893724 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-config\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.893765 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-combined-ca-bundle\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.893839 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.899480 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.900491 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.900803 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.919592 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1b956e3-372d-40f4-9f31-3028269f4175" path="/var/lib/kubelet/pods/f1b956e3-372d-40f4-9f31-3028269f4175/volumes" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.997336 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vwnh\" (UniqueName: \"kubernetes.io/projected/e2022f45-9ec9-483a-bce6-85945738da7f-kube-api-access-4vwnh\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.997650 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-config-data\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:58 crc kubenswrapper[4710]: I1002 13:07:58.998430 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9pjh\" (UniqueName: \"kubernetes.io/projected/f8748218-24af-4a2b-a6de-c7ea409ae6f8-kube-api-access-x9pjh\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.998807 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.998851 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-httpd-config\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.998893 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.998978 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-ovndb-tls-certs\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.999675 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-public-tls-certs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.999724 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8748218-24af-4a2b-a6de-c7ea409ae6f8-logs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.999787 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.999825 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-config\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:58.999872 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-combined-ca-bundle\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.007004 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-config\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.007575 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-httpd-config\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.014336 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-combined-ca-bundle\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.018861 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-ovndb-tls-certs\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.025857 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vwnh\" (UniqueName: \"kubernetes.io/projected/e2022f45-9ec9-483a-bce6-85945738da7f-kube-api-access-4vwnh\") pod \"neutron-f5fcfd868-8dmb9\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101288 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-config-data\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101360 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9pjh\" (UniqueName: \"kubernetes.io/projected/f8748218-24af-4a2b-a6de-c7ea409ae6f8-kube-api-access-x9pjh\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101405 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101430 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101465 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-public-tls-certs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101482 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8748218-24af-4a2b-a6de-c7ea409ae6f8-logs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.101501 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.107952 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8748218-24af-4a2b-a6de-c7ea409ae6f8-logs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.117387 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.118657 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-config-data\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.119410 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.121515 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-public-tls-certs\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.124288 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9pjh\" (UniqueName: \"kubernetes.io/projected/f8748218-24af-4a2b-a6de-c7ea409ae6f8-kube-api-access-x9pjh\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.124634 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8748218-24af-4a2b-a6de-c7ea409ae6f8-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"f8748218-24af-4a2b-a6de-c7ea409ae6f8\") " pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.144174 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.302064 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.381185 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.392459 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b4d46b6c-l4bbj"] Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.518669 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b868ba48-07b1-4a8e-8a4a-3c129840795c-logs\") pod \"b868ba48-07b1-4a8e-8a4a-3c129840795c\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.518819 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-combined-ca-bundle\") pod \"b868ba48-07b1-4a8e-8a4a-3c129840795c\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.518862 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-config-data\") pod \"b868ba48-07b1-4a8e-8a4a-3c129840795c\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.518940 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csrlp\" (UniqueName: \"kubernetes.io/projected/b868ba48-07b1-4a8e-8a4a-3c129840795c-kube-api-access-csrlp\") pod \"b868ba48-07b1-4a8e-8a4a-3c129840795c\" (UID: \"b868ba48-07b1-4a8e-8a4a-3c129840795c\") " Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.521059 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b868ba48-07b1-4a8e-8a4a-3c129840795c-logs" (OuterVolumeSpecName: "logs") pod "b868ba48-07b1-4a8e-8a4a-3c129840795c" (UID: "b868ba48-07b1-4a8e-8a4a-3c129840795c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.523643 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b868ba48-07b1-4a8e-8a4a-3c129840795c-kube-api-access-csrlp" (OuterVolumeSpecName: "kube-api-access-csrlp") pod "b868ba48-07b1-4a8e-8a4a-3c129840795c" (UID: "b868ba48-07b1-4a8e-8a4a-3c129840795c"). InnerVolumeSpecName "kube-api-access-csrlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.567717 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b868ba48-07b1-4a8e-8a4a-3c129840795c" (UID: "b868ba48-07b1-4a8e-8a4a-3c129840795c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.583289 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-config-data" (OuterVolumeSpecName: "config-data") pod "b868ba48-07b1-4a8e-8a4a-3c129840795c" (UID: "b868ba48-07b1-4a8e-8a4a-3c129840795c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.622947 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.623192 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b868ba48-07b1-4a8e-8a4a-3c129840795c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.623201 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csrlp\" (UniqueName: \"kubernetes.io/projected/b868ba48-07b1-4a8e-8a4a-3c129840795c-kube-api-access-csrlp\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.623214 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b868ba48-07b1-4a8e-8a4a-3c129840795c-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.653955 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.654003 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.675406 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" event={"ID":"48923995-d4ba-4562-8db2-26d5d34a0be5","Type":"ContainerStarted","Data":"94a1f583a0a32c8a1aec8fe50b8aecae0281a77bba4dec550e5ebd50a5b68360"} Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.678701 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"b868ba48-07b1-4a8e-8a4a-3c129840795c","Type":"ContainerDied","Data":"d8533f893ed916d916c08d9eb7ae19d2a5a85d67322ed63419519da0f68c2ae1"} Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.678737 4710 scope.go:117] "RemoveContainer" containerID="6d8a9f991e62c0267e565616f32771b9c7cf2afd1b3d07baa4c544315cc1f703" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.678857 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.701848 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m9cc" event={"ID":"6152e5af-f395-4bca-9aa1-19d17ec0ffa2","Type":"ContainerStarted","Data":"22317eba05ad5d22f364de32e34802eb9dddea15d9c3ac8122e08d32a81638fc"} Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.735812 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.770350 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.789473 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.844312 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.909502 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:59 crc kubenswrapper[4710]: E1002 13:07:59.909999 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b868ba48-07b1-4a8e-8a4a-3c129840795c" containerName="watcher-applier" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.910019 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b868ba48-07b1-4a8e-8a4a-3c129840795c" containerName="watcher-applier" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.910253 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b868ba48-07b1-4a8e-8a4a-3c129840795c" containerName="watcher-applier" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.911137 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.917861 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.935152 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:07:59 crc kubenswrapper[4710]: W1002 13:07:59.954953 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2022f45_9ec9_483a_bce6_85945738da7f.slice/crio-6f1aa040b5ac865f726204ed33a938f1af8a03bf1ae676bc109f26d14d9b9df7 WatchSource:0}: Error finding container 6f1aa040b5ac865f726204ed33a938f1af8a03bf1ae676bc109f26d14d9b9df7: Status 404 returned error can't find the container with id 6f1aa040b5ac865f726204ed33a938f1af8a03bf1ae676bc109f26d14d9b9df7 Oct 02 13:07:59 crc kubenswrapper[4710]: I1002 13:07:59.963217 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f5fcfd868-8dmb9"] Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.036582 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.051717 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfxgw\" (UniqueName: \"kubernetes.io/projected/3a69c931-f59d-4751-b560-a4b9f92cd3a2-kube-api-access-zfxgw\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.051827 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a69c931-f59d-4751-b560-a4b9f92cd3a2-config-data\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.051961 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a69c931-f59d-4751-b560-a4b9f92cd3a2-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.051987 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a69c931-f59d-4751-b560-a4b9f92cd3a2-logs\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.157358 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfxgw\" (UniqueName: \"kubernetes.io/projected/3a69c931-f59d-4751-b560-a4b9f92cd3a2-kube-api-access-zfxgw\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.157442 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a69c931-f59d-4751-b560-a4b9f92cd3a2-config-data\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.158228 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a69c931-f59d-4751-b560-a4b9f92cd3a2-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.158301 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a69c931-f59d-4751-b560-a4b9f92cd3a2-logs\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.158842 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a69c931-f59d-4751-b560-a4b9f92cd3a2-logs\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.162549 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a69c931-f59d-4751-b560-a4b9f92cd3a2-config-data\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.178474 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a69c931-f59d-4751-b560-a4b9f92cd3a2-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.182626 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfxgw\" (UniqueName: \"kubernetes.io/projected/3a69c931-f59d-4751-b560-a4b9f92cd3a2-kube-api-access-zfxgw\") pod \"watcher-applier-0\" (UID: \"3a69c931-f59d-4751-b560-a4b9f92cd3a2\") " pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.276266 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.726121 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f5fcfd868-8dmb9" event={"ID":"e2022f45-9ec9-483a-bce6-85945738da7f","Type":"ContainerStarted","Data":"30bbc857c2287e3dbf3e081f710f3c5e5f86bed6b01d689f0e337fbd7f3d1f1c"} Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.726833 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f5fcfd868-8dmb9" event={"ID":"e2022f45-9ec9-483a-bce6-85945738da7f","Type":"ContainerStarted","Data":"6f1aa040b5ac865f726204ed33a938f1af8a03bf1ae676bc109f26d14d9b9df7"} Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.735655 4710 generic.go:334] "Generic (PLEG): container finished" podID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerID="4e136af53a5996f4f900357b870ceca06f7437fbf36fb4667abd6c76f1158937" exitCode=0 Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.735770 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" event={"ID":"48923995-d4ba-4562-8db2-26d5d34a0be5","Type":"ContainerDied","Data":"4e136af53a5996f4f900357b870ceca06f7437fbf36fb4667abd6c76f1158937"} Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.741003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f8748218-24af-4a2b-a6de-c7ea409ae6f8","Type":"ContainerStarted","Data":"c7fbe440536df10e03d6640b937ee4521379c87adfde6413c336f7eee915b1c3"} Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.741228 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f8748218-24af-4a2b-a6de-c7ea409ae6f8","Type":"ContainerStarted","Data":"f090bbfe057a310294775fe3c8148ff21a8e69d828c0c85110fdc553d465b497"} Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.742808 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.742852 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.786301 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4m9cc" podStartSLOduration=6.860633698 podStartE2EDuration="49.786281295s" podCreationTimestamp="2025-10-02 13:07:11 +0000 UTC" firstStartedPulling="2025-10-02 13:07:14.126849108 +0000 UTC m=+1078.233259991" lastFinishedPulling="2025-10-02 13:07:57.052496705 +0000 UTC m=+1121.158907588" observedRunningTime="2025-10-02 13:08:00.779292139 +0000 UTC m=+1124.885703022" watchObservedRunningTime="2025-10-02 13:08:00.786281295 +0000 UTC m=+1124.892692178" Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.844001 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 13:08:00 crc kubenswrapper[4710]: I1002 13:08:00.922063 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b868ba48-07b1-4a8e-8a4a-3c129840795c" path="/var/lib/kubelet/pods/b868ba48-07b1-4a8e-8a4a-3c129840795c/volumes" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.030919 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c89fc49-c8w65"] Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.038064 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.041528 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.041778 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.068341 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c89fc49-c8w65"] Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.149933 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.164:8443: connect: connection refused" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.191953 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-config\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.192183 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2st2t\" (UniqueName: \"kubernetes.io/projected/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-kube-api-access-2st2t\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.192221 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-httpd-config\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.192273 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-ovndb-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.192298 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-internal-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.192378 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-public-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.192402 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-combined-ca-bundle\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293733 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-public-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293798 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-combined-ca-bundle\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293846 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-config\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293869 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2st2t\" (UniqueName: \"kubernetes.io/projected/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-kube-api-access-2st2t\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293896 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-httpd-config\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293943 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-ovndb-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.293963 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-internal-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.300438 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-ovndb-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.300690 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-config\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.300696 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-combined-ca-bundle\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.303455 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-public-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.303665 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-httpd-config\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.306355 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-internal-tls-certs\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.314041 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2st2t\" (UniqueName: \"kubernetes.io/projected/cf6923e6-0cef-430d-bf95-e6d3aa1a30e4-kube-api-access-2st2t\") pod \"neutron-6c89fc49-c8w65\" (UID: \"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4\") " pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.317956 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-58474b8ddd-nzgx9" podUID="b8267ea9-4756-45d5-a7bc-8985275cf297" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.165:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.165:8443: connect: connection refused" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.427425 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.767044 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"3a69c931-f59d-4751-b560-a4b9f92cd3a2","Type":"ContainerStarted","Data":"f0fccb1b5db71ec17d993bda8fb97471b8813e8233e4d588e174e73e24455a93"} Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.767633 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"3a69c931-f59d-4751-b560-a4b9f92cd3a2","Type":"ContainerStarted","Data":"a76c83ec6d26227998d67a32a37007ec4ff015a2eee01b75e8b211d0d6b45143"} Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.789017 4710 generic.go:334] "Generic (PLEG): container finished" podID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerID="12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3" exitCode=1 Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.789078 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerDied","Data":"12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3"} Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.789109 4710 scope.go:117] "RemoveContainer" containerID="c65aefbb3a54f1ff3f30d8267f4fb52935f8b31fe351a22c4c0e9518a26b77cf" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.789692 4710 scope.go:117] "RemoveContainer" containerID="12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3" Oct 02 13:08:01 crc kubenswrapper[4710]: E1002 13:08:01.797316 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.799043 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.799018446 podStartE2EDuration="2.799018446s" podCreationTimestamp="2025-10-02 13:07:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:01.788055661 +0000 UTC m=+1125.894466544" watchObservedRunningTime="2025-10-02 13:08:01.799018446 +0000 UTC m=+1125.905429339" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.817555 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" event={"ID":"48923995-d4ba-4562-8db2-26d5d34a0be5","Type":"ContainerStarted","Data":"276ed2623050743a03d0a90e28cb257b01449d397ca1836d8531fbe9afd03248"} Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.818604 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.826279 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"f8748218-24af-4a2b-a6de-c7ea409ae6f8","Type":"ContainerStarted","Data":"b3c1a802855b140ed19f287b4ffed41be4ae47f4cebebf5a792c09c7ca7a41ef"} Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.827286 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.830738 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f5fcfd868-8dmb9" event={"ID":"e2022f45-9ec9-483a-bce6-85945738da7f","Type":"ContainerStarted","Data":"ad530a2586f341efcb8a07c56a3915e8efc95ca122fbbd798351b9a66fa58c45"} Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.831074 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.884780 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" podStartSLOduration=3.884731994 podStartE2EDuration="3.884731994s" podCreationTimestamp="2025-10-02 13:07:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:01.873219884 +0000 UTC m=+1125.979630767" watchObservedRunningTime="2025-10-02 13:08:01.884731994 +0000 UTC m=+1125.991142877" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.900706 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f5fcfd868-8dmb9" podStartSLOduration=3.900683885 podStartE2EDuration="3.900683885s" podCreationTimestamp="2025-10-02 13:07:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:01.887890143 +0000 UTC m=+1125.994301026" watchObservedRunningTime="2025-10-02 13:08:01.900683885 +0000 UTC m=+1126.007094768" Oct 02 13:08:01 crc kubenswrapper[4710]: I1002 13:08:01.929735 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=3.929716576 podStartE2EDuration="3.929716576s" podCreationTimestamp="2025-10-02 13:07:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:01.919569771 +0000 UTC m=+1126.025980674" watchObservedRunningTime="2025-10-02 13:08:01.929716576 +0000 UTC m=+1126.036127459" Oct 02 13:08:02 crc kubenswrapper[4710]: I1002 13:08:02.093476 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:02 crc kubenswrapper[4710]: I1002 13:08:02.093559 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:02 crc kubenswrapper[4710]: I1002 13:08:02.185609 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c89fc49-c8w65"] Oct 02 13:08:02 crc kubenswrapper[4710]: I1002 13:08:02.847825 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c89fc49-c8w65" event={"ID":"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4","Type":"ContainerStarted","Data":"1e10c86eccbfce0ea58d35146cb35bc96913d9407a82fa999ea233d69d0ff205"} Oct 02 13:08:02 crc kubenswrapper[4710]: I1002 13:08:02.855633 4710 scope.go:117] "RemoveContainer" containerID="12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3" Oct 02 13:08:02 crc kubenswrapper[4710]: E1002 13:08:02.855951 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:03 crc kubenswrapper[4710]: I1002 13:08:03.279052 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 13:08:03 crc kubenswrapper[4710]: I1002 13:08:03.872434 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c89fc49-c8w65" event={"ID":"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4","Type":"ContainerStarted","Data":"0a296e0472d249181ca2952c816cdc484a2116b42604fdcdddfa4b6b109d011f"} Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.302939 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.303039 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.698563 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.709065 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.884885 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c89fc49-c8w65" event={"ID":"cf6923e6-0cef-430d-bf95-e6d3aa1a30e4","Type":"ContainerStarted","Data":"c445642f9b13316de1f4937e8a04dcb92f5c81a43cc9fcd4a2ade4a33290dc89"} Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.885211 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:04 crc kubenswrapper[4710]: I1002 13:08:04.914478 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c89fc49-c8w65" podStartSLOduration=4.914459943 podStartE2EDuration="4.914459943s" podCreationTimestamp="2025-10-02 13:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:04.907874057 +0000 UTC m=+1129.014284960" watchObservedRunningTime="2025-10-02 13:08:04.914459943 +0000 UTC m=+1129.020870816" Oct 02 13:08:05 crc kubenswrapper[4710]: I1002 13:08:05.000680 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 13:08:05 crc kubenswrapper[4710]: I1002 13:08:05.277422 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 02 13:08:05 crc kubenswrapper[4710]: I1002 13:08:05.454017 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 13:08:08 crc kubenswrapper[4710]: I1002 13:08:08.884008 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:08:08 crc kubenswrapper[4710]: I1002 13:08:08.944257 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849796779c-f5465"] Oct 02 13:08:08 crc kubenswrapper[4710]: I1002 13:08:08.944515 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-849796779c-f5465" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="dnsmasq-dns" containerID="cri-o://d363b7e81e22cd40e6805e7eb70d6bbee2e985e344a8ad40a9a201cff48832f9" gracePeriod=10 Oct 02 13:08:09 crc kubenswrapper[4710]: I1002 13:08:09.302865 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 02 13:08:09 crc kubenswrapper[4710]: I1002 13:08:09.313364 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 02 13:08:09 crc kubenswrapper[4710]: I1002 13:08:09.940858 4710 generic.go:334] "Generic (PLEG): container finished" podID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerID="d363b7e81e22cd40e6805e7eb70d6bbee2e985e344a8ad40a9a201cff48832f9" exitCode=0 Oct 02 13:08:09 crc kubenswrapper[4710]: I1002 13:08:09.941148 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849796779c-f5465" event={"ID":"6d28154a-1fc9-4d45-bb67-ecac912c5d01","Type":"ContainerDied","Data":"d363b7e81e22cd40e6805e7eb70d6bbee2e985e344a8ad40a9a201cff48832f9"} Oct 02 13:08:09 crc kubenswrapper[4710]: I1002 13:08:09.952998 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 13:08:10 crc kubenswrapper[4710]: I1002 13:08:10.277515 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 02 13:08:10 crc kubenswrapper[4710]: I1002 13:08:10.310214 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 02 13:08:10 crc kubenswrapper[4710]: I1002 13:08:10.984604 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 02 13:08:12 crc kubenswrapper[4710]: I1002 13:08:12.092293 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:12 crc kubenswrapper[4710]: I1002 13:08:12.093374 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:12 crc kubenswrapper[4710]: I1002 13:08:12.094304 4710 scope.go:117] "RemoveContainer" containerID="12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3" Oct 02 13:08:12 crc kubenswrapper[4710]: I1002 13:08:12.406272 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-849796779c-f5465" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.639060 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.644583 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.808392 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:08:13 crc kubenswrapper[4710]: E1002 13:08:13.910274 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.972966 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-nb\") pod \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.973013 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-svc\") pod \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.977028 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-swift-storage-0\") pod \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.977089 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-config\") pod \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.977119 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-sb\") pod \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.977152 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfr8t\" (UniqueName: \"kubernetes.io/projected/6d28154a-1fc9-4d45-bb67-ecac912c5d01-kube-api-access-zfr8t\") pod \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\" (UID: \"6d28154a-1fc9-4d45-bb67-ecac912c5d01\") " Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.986295 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d28154a-1fc9-4d45-bb67-ecac912c5d01-kube-api-access-zfr8t" (OuterVolumeSpecName: "kube-api-access-zfr8t") pod "6d28154a-1fc9-4d45-bb67-ecac912c5d01" (UID: "6d28154a-1fc9-4d45-bb67-ecac912c5d01"). InnerVolumeSpecName "kube-api-access-zfr8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.992211 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849796779c-f5465" Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.992199 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849796779c-f5465" event={"ID":"6d28154a-1fc9-4d45-bb67-ecac912c5d01","Type":"ContainerDied","Data":"7d2bed6ee2ae2b48b30fc3bef8960567e1132db5820fbad777c2fd0d7084b629"} Oct 02 13:08:13 crc kubenswrapper[4710]: I1002 13:08:13.992536 4710 scope.go:117] "RemoveContainer" containerID="d363b7e81e22cd40e6805e7eb70d6bbee2e985e344a8ad40a9a201cff48832f9" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.003131 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerStarted","Data":"341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec"} Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.017464 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerStarted","Data":"ec2e13a9bfec1edcca0d4a80db652e7b9929f4e2b1328f686fedcb3856ec00a3"} Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.017631 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="ceilometer-notification-agent" containerID="cri-o://1dc130c3972d51c15bd9ed2746f526598c726a4f4dcc16e5440157a94ad252e1" gracePeriod=30 Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.018004 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.018069 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="proxy-httpd" containerID="cri-o://ec2e13a9bfec1edcca0d4a80db652e7b9929f4e2b1328f686fedcb3856ec00a3" gracePeriod=30 Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.018144 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="sg-core" containerID="cri-o://f31d2c3b09a941fffe2619fe6a392f8621451418117fb3f7673b22acebccc053" gracePeriod=30 Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.029979 4710 scope.go:117] "RemoveContainer" containerID="ddfaea0fa96c0d46a5320c401e70052029af0e813a774b8515bc1dfc32ec0340" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.051260 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6d28154a-1fc9-4d45-bb67-ecac912c5d01" (UID: "6d28154a-1fc9-4d45-bb67-ecac912c5d01"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.052699 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6d28154a-1fc9-4d45-bb67-ecac912c5d01" (UID: "6d28154a-1fc9-4d45-bb67-ecac912c5d01"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.063206 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6d28154a-1fc9-4d45-bb67-ecac912c5d01" (UID: "6d28154a-1fc9-4d45-bb67-ecac912c5d01"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.063549 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6d28154a-1fc9-4d45-bb67-ecac912c5d01" (UID: "6d28154a-1fc9-4d45-bb67-ecac912c5d01"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.073826 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-config" (OuterVolumeSpecName: "config") pod "6d28154a-1fc9-4d45-bb67-ecac912c5d01" (UID: "6d28154a-1fc9-4d45-bb67-ecac912c5d01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.080790 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.080823 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.080832 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.080841 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfr8t\" (UniqueName: \"kubernetes.io/projected/6d28154a-1fc9-4d45-bb67-ecac912c5d01-kube-api-access-zfr8t\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.080851 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.080860 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d28154a-1fc9-4d45-bb67-ecac912c5d01-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.327686 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849796779c-f5465"] Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.336562 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-849796779c-f5465"] Oct 02 13:08:14 crc kubenswrapper[4710]: I1002 13:08:14.918429 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" path="/var/lib/kubelet/pods/6d28154a-1fc9-4d45-bb67-ecac912c5d01/volumes" Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.031100 4710 generic.go:334] "Generic (PLEG): container finished" podID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerID="ec2e13a9bfec1edcca0d4a80db652e7b9929f4e2b1328f686fedcb3856ec00a3" exitCode=0 Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.031387 4710 generic.go:334] "Generic (PLEG): container finished" podID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerID="f31d2c3b09a941fffe2619fe6a392f8621451418117fb3f7673b22acebccc053" exitCode=2 Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.032296 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerDied","Data":"ec2e13a9bfec1edcca0d4a80db652e7b9929f4e2b1328f686fedcb3856ec00a3"} Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.032427 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerDied","Data":"f31d2c3b09a941fffe2619fe6a392f8621451418117fb3f7673b22acebccc053"} Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.575447 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-58474b8ddd-nzgx9" Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.646996 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f85867556-bdslh"] Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.648233 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon-log" containerID="cri-o://a42ca0c2045d5f637fdd39fdecb14ac1b0dd3e6fe65cc97a6a353834d22c6f67" gracePeriod=30 Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.648532 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" containerID="cri-o://2b4a4bb8893f083849330b9ab1404fe925487d732359bb18512233446d936a8b" gracePeriod=30 Oct 02 13:08:15 crc kubenswrapper[4710]: I1002 13:08:15.673028 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.059988 4710 generic.go:334] "Generic (PLEG): container finished" podID="6b39391f-b6e4-4275-941b-63484ea90c69" containerID="2a6b248d51fa9075d87a5aba4358c9b232b87f161558f849071c0ec412f1348a" exitCode=137 Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.060243 4710 generic.go:334] "Generic (PLEG): container finished" podID="6b39391f-b6e4-4275-941b-63484ea90c69" containerID="491cc9b96d286224c8c387787f38c76742c3e9412da4f7ad4fec674adbc334a7" exitCode=137 Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.060265 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-588846bfb9-5mhkc" event={"ID":"6b39391f-b6e4-4275-941b-63484ea90c69","Type":"ContainerDied","Data":"2a6b248d51fa9075d87a5aba4358c9b232b87f161558f849071c0ec412f1348a"} Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.060289 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-588846bfb9-5mhkc" event={"ID":"6b39391f-b6e4-4275-941b-63484ea90c69","Type":"ContainerDied","Data":"491cc9b96d286224c8c387787f38c76742c3e9412da4f7ad4fec674adbc334a7"} Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.271693 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.463928 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-config-data\") pod \"6b39391f-b6e4-4275-941b-63484ea90c69\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.464155 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-scripts\") pod \"6b39391f-b6e4-4275-941b-63484ea90c69\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.464287 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b39391f-b6e4-4275-941b-63484ea90c69-horizon-secret-key\") pod \"6b39391f-b6e4-4275-941b-63484ea90c69\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.464344 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rtkd\" (UniqueName: \"kubernetes.io/projected/6b39391f-b6e4-4275-941b-63484ea90c69-kube-api-access-8rtkd\") pod \"6b39391f-b6e4-4275-941b-63484ea90c69\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.464491 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b39391f-b6e4-4275-941b-63484ea90c69-logs\") pod \"6b39391f-b6e4-4275-941b-63484ea90c69\" (UID: \"6b39391f-b6e4-4275-941b-63484ea90c69\") " Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.465074 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b39391f-b6e4-4275-941b-63484ea90c69-logs" (OuterVolumeSpecName: "logs") pod "6b39391f-b6e4-4275-941b-63484ea90c69" (UID: "6b39391f-b6e4-4275-941b-63484ea90c69"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.478010 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b39391f-b6e4-4275-941b-63484ea90c69-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6b39391f-b6e4-4275-941b-63484ea90c69" (UID: "6b39391f-b6e4-4275-941b-63484ea90c69"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.478085 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b39391f-b6e4-4275-941b-63484ea90c69-kube-api-access-8rtkd" (OuterVolumeSpecName: "kube-api-access-8rtkd") pod "6b39391f-b6e4-4275-941b-63484ea90c69" (UID: "6b39391f-b6e4-4275-941b-63484ea90c69"). InnerVolumeSpecName "kube-api-access-8rtkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.500601 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-config-data" (OuterVolumeSpecName: "config-data") pod "6b39391f-b6e4-4275-941b-63484ea90c69" (UID: "6b39391f-b6e4-4275-941b-63484ea90c69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.500888 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-scripts" (OuterVolumeSpecName: "scripts") pod "6b39391f-b6e4-4275-941b-63484ea90c69" (UID: "6b39391f-b6e4-4275-941b-63484ea90c69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.566889 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b39391f-b6e4-4275-941b-63484ea90c69-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.566927 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.566939 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b39391f-b6e4-4275-941b-63484ea90c69-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.566950 4710 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b39391f-b6e4-4275-941b-63484ea90c69-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.566967 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rtkd\" (UniqueName: \"kubernetes.io/projected/6b39391f-b6e4-4275-941b-63484ea90c69-kube-api-access-8rtkd\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:16 crc kubenswrapper[4710]: I1002 13:08:16.779515 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:59708->10.217.0.164:8443: read: connection reset by peer" Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.072287 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-588846bfb9-5mhkc" Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.072274 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-588846bfb9-5mhkc" event={"ID":"6b39391f-b6e4-4275-941b-63484ea90c69","Type":"ContainerDied","Data":"3aa5b085b06abaff31c452b9924bbf80c23284695fdd63edd1eca89839e4ddf9"} Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.072503 4710 scope.go:117] "RemoveContainer" containerID="2a6b248d51fa9075d87a5aba4358c9b232b87f161558f849071c0ec412f1348a" Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.074435 4710 generic.go:334] "Generic (PLEG): container finished" podID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerID="2b4a4bb8893f083849330b9ab1404fe925487d732359bb18512233446d936a8b" exitCode=0 Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.074466 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f85867556-bdslh" event={"ID":"00020afa-1e01-4a2c-924e-2aa55bac0e85","Type":"ContainerDied","Data":"2b4a4bb8893f083849330b9ab1404fe925487d732359bb18512233446d936a8b"} Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.102360 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-588846bfb9-5mhkc"] Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.114713 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-588846bfb9-5mhkc"] Oct 02 13:08:17 crc kubenswrapper[4710]: I1002 13:08:17.246639 4710 scope.go:117] "RemoveContainer" containerID="491cc9b96d286224c8c387787f38c76742c3e9412da4f7ad4fec674adbc334a7" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.090103 4710 generic.go:334] "Generic (PLEG): container finished" podID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerID="1dc130c3972d51c15bd9ed2746f526598c726a4f4dcc16e5440157a94ad252e1" exitCode=0 Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.090189 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerDied","Data":"1dc130c3972d51c15bd9ed2746f526598c726a4f4dcc16e5440157a94ad252e1"} Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.090698 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aed5f89f-6790-4f1f-9ed2-2f075f8d4647","Type":"ContainerDied","Data":"2d484683f04090875bc64dae71a5cd9d7078a47a72bc1956266bcc3edf2b6691"} Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.090718 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d484683f04090875bc64dae71a5cd9d7078a47a72bc1956266bcc3edf2b6691" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.099272 4710 generic.go:334] "Generic (PLEG): container finished" podID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerID="341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec" exitCode=1 Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.099314 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerDied","Data":"341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec"} Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.099351 4710 scope.go:117] "RemoveContainer" containerID="12c1ae21b237f2ce8ddb9c2ae74dcf01985a92a8eb8b4bb65bf4b13743bc08a3" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.099839 4710 scope.go:117] "RemoveContainer" containerID="341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec" Oct 02 13:08:18 crc kubenswrapper[4710]: E1002 13:08:18.100209 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.180298 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197424 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-sg-core-conf-yaml\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197474 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-run-httpd\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197535 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-config-data\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197569 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-log-httpd\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197604 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crqtp\" (UniqueName: \"kubernetes.io/projected/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-kube-api-access-crqtp\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197637 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-combined-ca-bundle\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.197686 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-scripts\") pod \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\" (UID: \"aed5f89f-6790-4f1f-9ed2-2f075f8d4647\") " Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.198085 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.208868 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.212637 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-scripts" (OuterVolumeSpecName: "scripts") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.227863 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-kube-api-access-crqtp" (OuterVolumeSpecName: "kube-api-access-crqtp") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "kube-api-access-crqtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.244427 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.252155 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.300425 4710 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.300461 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crqtp\" (UniqueName: \"kubernetes.io/projected/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-kube-api-access-crqtp\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.300476 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.300487 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.300497 4710 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.300508 4710 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.301931 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-config-data" (OuterVolumeSpecName: "config-data") pod "aed5f89f-6790-4f1f-9ed2-2f075f8d4647" (UID: "aed5f89f-6790-4f1f-9ed2-2f075f8d4647"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.402624 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5f89f-6790-4f1f-9ed2-2f075f8d4647-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:18 crc kubenswrapper[4710]: I1002 13:08:18.921522 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" path="/var/lib/kubelet/pods/6b39391f-b6e4-4275-941b-63484ea90c69/volumes" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.112265 4710 generic.go:334] "Generic (PLEG): container finished" podID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" containerID="712b61b517d05149b02904aa2054f336915e1e5e34db2dae2941851eac765724" exitCode=0 Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.112653 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.113495 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nwcvq" event={"ID":"f53ffe3d-d27a-41dd-a98b-4debca318ec4","Type":"ContainerDied","Data":"712b61b517d05149b02904aa2054f336915e1e5e34db2dae2941851eac765724"} Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.194428 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.206892 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220141 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220596 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="ceilometer-notification-agent" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220620 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="ceilometer-notification-agent" Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220652 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="dnsmasq-dns" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220660 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="dnsmasq-dns" Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220679 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="sg-core" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220686 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="sg-core" Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220704 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon-log" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220711 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon-log" Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220718 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="init" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220726 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="init" Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220767 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="proxy-httpd" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220774 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="proxy-httpd" Oct 02 13:08:19 crc kubenswrapper[4710]: E1002 13:08:19.220783 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.220791 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.221007 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.221018 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d28154a-1fc9-4d45-bb67-ecac912c5d01" containerName="dnsmasq-dns" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.221030 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="sg-core" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.221048 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b39391f-b6e4-4275-941b-63484ea90c69" containerName="horizon-log" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.221060 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="ceilometer-notification-agent" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.221076 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" containerName="proxy-httpd" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.223146 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.226182 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.226385 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.230016 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423069 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-run-httpd\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423133 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-log-httpd\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423315 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-scripts\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423414 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423447 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-config-data\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423476 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.423552 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b96vg\" (UniqueName: \"kubernetes.io/projected/7dc3ecad-515a-40b5-ac68-b19c0d276caa-kube-api-access-b96vg\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.524973 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b96vg\" (UniqueName: \"kubernetes.io/projected/7dc3ecad-515a-40b5-ac68-b19c0d276caa-kube-api-access-b96vg\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.525074 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-run-httpd\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.525115 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-log-httpd\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.525189 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-scripts\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.525224 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.525256 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-config-data\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.525280 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.526069 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-run-httpd\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.526447 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-log-httpd\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.529930 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-scripts\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.529967 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-config-data\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.530482 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.533047 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.547567 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b96vg\" (UniqueName: \"kubernetes.io/projected/7dc3ecad-515a-40b5-ac68-b19c0d276caa-kube-api-access-b96vg\") pod \"ceilometer-0\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " pod="openstack/ceilometer-0" Oct 02 13:08:19 crc kubenswrapper[4710]: I1002 13:08:19.581899 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.020832 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.125027 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerStarted","Data":"244a92e59f11e6a642c2aa8e7b8181582366221f843e6a39f5052812d24568ca"} Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.757883 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.871165 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-db-sync-config-data\") pod \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.871295 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkj2j\" (UniqueName: \"kubernetes.io/projected/f53ffe3d-d27a-41dd-a98b-4debca318ec4-kube-api-access-pkj2j\") pod \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.871369 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-combined-ca-bundle\") pod \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\" (UID: \"f53ffe3d-d27a-41dd-a98b-4debca318ec4\") " Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.877262 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f53ffe3d-d27a-41dd-a98b-4debca318ec4" (UID: "f53ffe3d-d27a-41dd-a98b-4debca318ec4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.882998 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f53ffe3d-d27a-41dd-a98b-4debca318ec4-kube-api-access-pkj2j" (OuterVolumeSpecName: "kube-api-access-pkj2j") pod "f53ffe3d-d27a-41dd-a98b-4debca318ec4" (UID: "f53ffe3d-d27a-41dd-a98b-4debca318ec4"). InnerVolumeSpecName "kube-api-access-pkj2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.903893 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f53ffe3d-d27a-41dd-a98b-4debca318ec4" (UID: "f53ffe3d-d27a-41dd-a98b-4debca318ec4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.923127 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed5f89f-6790-4f1f-9ed2-2f075f8d4647" path="/var/lib/kubelet/pods/aed5f89f-6790-4f1f-9ed2-2f075f8d4647/volumes" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.973618 4710 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.973665 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkj2j\" (UniqueName: \"kubernetes.io/projected/f53ffe3d-d27a-41dd-a98b-4debca318ec4-kube-api-access-pkj2j\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:20 crc kubenswrapper[4710]: I1002 13:08:20.973681 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53ffe3d-d27a-41dd-a98b-4debca318ec4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.136261 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.164:8443: connect: connection refused" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.141496 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nwcvq" event={"ID":"f53ffe3d-d27a-41dd-a98b-4debca318ec4","Type":"ContainerDied","Data":"1f7fdfba007e7488b49e275a0c9a3ad53eaca5a6c22e531e45bfa32e377b3ca6"} Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.141538 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f7fdfba007e7488b49e275a0c9a3ad53eaca5a6c22e531e45bfa32e377b3ca6" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.141601 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nwcvq" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.145412 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerStarted","Data":"a7572dc21b0e3541f4fd4ee817688f46e16ed0b9c70d512f30df4b8982ca8772"} Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.145452 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerStarted","Data":"b4d999f5769fbff464e6d35ac58c689edf5e54b43e48465dea610111f633ef4e"} Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.442635 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5c86c47bf8-5k59q"] Oct 02 13:08:21 crc kubenswrapper[4710]: E1002 13:08:21.443184 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" containerName="barbican-db-sync" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.443205 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" containerName="barbican-db-sync" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.443407 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" containerName="barbican-db-sync" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.444555 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.454094 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-f7dcd7d87-c6j6p"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.454540 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.454949 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.455268 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gq2s5" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.455938 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.478120 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.501017 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-combined-ca-bundle\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.501115 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-config-data-custom\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.501218 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddcb91c7-f839-4756-9714-bf04aaf2382a-logs\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.501345 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5v8m\" (UniqueName: \"kubernetes.io/projected/ddcb91c7-f839-4756-9714-bf04aaf2382a-kube-api-access-p5v8m\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.501401 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-config-data\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.503115 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c86c47bf8-5k59q"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.577471 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f7dcd7d87-c6j6p"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610583 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593ff4da-0e41-460e-9b1c-b8d0d032d492-logs\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610666 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5v8m\" (UniqueName: \"kubernetes.io/projected/ddcb91c7-f839-4756-9714-bf04aaf2382a-kube-api-access-p5v8m\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610711 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-config-data\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610838 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-combined-ca-bundle\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610891 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-config-data-custom\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610946 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-combined-ca-bundle\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.610989 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-config-data-custom\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.611035 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-config-data\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.611068 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2lgq\" (UniqueName: \"kubernetes.io/projected/593ff4da-0e41-460e-9b1c-b8d0d032d492-kube-api-access-q2lgq\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.611106 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddcb91c7-f839-4756-9714-bf04aaf2382a-logs\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.616225 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddcb91c7-f839-4756-9714-bf04aaf2382a-logs\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.622380 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c8bffbc5c-kxntj"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.626189 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-config-data\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.626794 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.638183 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-config-data-custom\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.640029 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddcb91c7-f839-4756-9714-bf04aaf2382a-combined-ca-bundle\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.670189 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5v8m\" (UniqueName: \"kubernetes.io/projected/ddcb91c7-f839-4756-9714-bf04aaf2382a-kube-api-access-p5v8m\") pod \"barbican-worker-f7dcd7d87-c6j6p\" (UID: \"ddcb91c7-f839-4756-9714-bf04aaf2382a\") " pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.674018 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8bffbc5c-kxntj"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.686761 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-844bdb96b6-wthf8"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.689476 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.694498 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.705819 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-844bdb96b6-wthf8"] Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.712281 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-config-data\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.712592 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2lgq\" (UniqueName: \"kubernetes.io/projected/593ff4da-0e41-460e-9b1c-b8d0d032d492-kube-api-access-q2lgq\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.712652 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593ff4da-0e41-460e-9b1c-b8d0d032d492-logs\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.712774 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-combined-ca-bundle\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.712812 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-config-data-custom\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.718412 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593ff4da-0e41-460e-9b1c-b8d0d032d492-logs\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.733518 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-combined-ca-bundle\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.738880 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-config-data\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.739529 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/593ff4da-0e41-460e-9b1c-b8d0d032d492-config-data-custom\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.759314 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2lgq\" (UniqueName: \"kubernetes.io/projected/593ff4da-0e41-460e-9b1c-b8d0d032d492-kube-api-access-q2lgq\") pod \"barbican-keystone-listener-5c86c47bf8-5k59q\" (UID: \"593ff4da-0e41-460e-9b1c-b8d0d032d492\") " pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.819420 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.819580 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6931f-1351-4285-a8fc-e1280f21dee6-logs\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.819666 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpkrl\" (UniqueName: \"kubernetes.io/projected/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-kube-api-access-wpkrl\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.819915 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-svc\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.819975 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-config\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.820067 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.820184 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.820292 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7g5h\" (UniqueName: \"kubernetes.io/projected/66d6931f-1351-4285-a8fc-e1280f21dee6-kube-api-access-h7g5h\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.820363 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data-custom\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.820781 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-combined-ca-bundle\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.821085 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.834867 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.907231 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f7dcd7d87-c6j6p" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.925714 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.925811 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7g5h\" (UniqueName: \"kubernetes.io/projected/66d6931f-1351-4285-a8fc-e1280f21dee6-kube-api-access-h7g5h\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.925855 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data-custom\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.925915 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-combined-ca-bundle\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926010 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926037 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926065 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6931f-1351-4285-a8fc-e1280f21dee6-logs\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926100 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpkrl\" (UniqueName: \"kubernetes.io/projected/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-kube-api-access-wpkrl\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926171 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-svc\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926200 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-config\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.926255 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.927491 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.928435 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.930333 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6931f-1351-4285-a8fc-e1280f21dee6-logs\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.930931 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-svc\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.930996 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-config\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.931861 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.937683 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.950879 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-combined-ca-bundle\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.953511 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data-custom\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.955234 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7g5h\" (UniqueName: \"kubernetes.io/projected/66d6931f-1351-4285-a8fc-e1280f21dee6-kube-api-access-h7g5h\") pod \"barbican-api-844bdb96b6-wthf8\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:21 crc kubenswrapper[4710]: I1002 13:08:21.956734 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpkrl\" (UniqueName: \"kubernetes.io/projected/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-kube-api-access-wpkrl\") pod \"dnsmasq-dns-5c8bffbc5c-kxntj\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.094050 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.094467 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.095354 4710 scope.go:117] "RemoveContainer" containerID="341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec" Oct 02 13:08:22 crc kubenswrapper[4710]: E1002 13:08:22.095702 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.165422 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.173821 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m9cc" event={"ID":"6152e5af-f395-4bca-9aa1-19d17ec0ffa2","Type":"ContainerDied","Data":"22317eba05ad5d22f364de32e34802eb9dddea15d9c3ac8122e08d32a81638fc"} Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.173702 4710 generic.go:334] "Generic (PLEG): container finished" podID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" containerID="22317eba05ad5d22f364de32e34802eb9dddea15d9c3ac8122e08d32a81638fc" exitCode=0 Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.179280 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.410209 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c86c47bf8-5k59q"] Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.530367 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.530427 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.631968 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f7dcd7d87-c6j6p"] Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.854934 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8bffbc5c-kxntj"] Oct 02 13:08:22 crc kubenswrapper[4710]: I1002 13:08:22.871046 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-844bdb96b6-wthf8"] Oct 02 13:08:23 crc kubenswrapper[4710]: I1002 13:08:23.192885 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" event={"ID":"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd","Type":"ContainerStarted","Data":"8abecb6ed1e69c5cbedeaee15282ef35b18954fabb442fda22d92be7861ffb96"} Oct 02 13:08:23 crc kubenswrapper[4710]: I1002 13:08:23.205028 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" event={"ID":"593ff4da-0e41-460e-9b1c-b8d0d032d492","Type":"ContainerStarted","Data":"6fe36b26ceab457ab06096e8635193ed3ba90b2024017310527a360133d540a1"} Oct 02 13:08:23 crc kubenswrapper[4710]: I1002 13:08:23.209819 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-844bdb96b6-wthf8" event={"ID":"66d6931f-1351-4285-a8fc-e1280f21dee6","Type":"ContainerStarted","Data":"930bb7269bce50194d6b64f6d3c2da70c70872110c60a9469e2d65b982dcb152"} Oct 02 13:08:23 crc kubenswrapper[4710]: I1002 13:08:23.217132 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f7dcd7d87-c6j6p" event={"ID":"ddcb91c7-f839-4756-9714-bf04aaf2382a","Type":"ContainerStarted","Data":"5d480de95eebe4df1906c9da82f2fa7448d79469375102bef919f783c8db132a"} Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.255388 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-844bdb96b6-wthf8" event={"ID":"66d6931f-1351-4285-a8fc-e1280f21dee6","Type":"ContainerStarted","Data":"28bcdefa12b42f75cfb4313625fc0a6c9ee02596c8f22e5d194091d3ff8d682f"} Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.255759 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-844bdb96b6-wthf8" event={"ID":"66d6931f-1351-4285-a8fc-e1280f21dee6","Type":"ContainerStarted","Data":"5fd49cd51631f5f283cd0ad640e6fe5d344ff2ea3219e3b57ecb8c2cc2517c4b"} Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.255785 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.255800 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.259099 4710 generic.go:334] "Generic (PLEG): container finished" podID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerID="63d431214dd89183d41cf9cbf34267645fbd1fd0706bf82bb600375ca2c3512e" exitCode=0 Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.259139 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" event={"ID":"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd","Type":"ContainerDied","Data":"63d431214dd89183d41cf9cbf34267645fbd1fd0706bf82bb600375ca2c3512e"} Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.286178 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-844bdb96b6-wthf8" podStartSLOduration=3.286154713 podStartE2EDuration="3.286154713s" podCreationTimestamp="2025-10-02 13:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:24.276388377 +0000 UTC m=+1148.382799270" watchObservedRunningTime="2025-10-02 13:08:24.286154713 +0000 UTC m=+1148.392565596" Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.953883 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54f666d7d6-zzpk4"] Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.955855 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.959988 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.962548 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 13:08:24 crc kubenswrapper[4710]: I1002 13:08:24.970257 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54f666d7d6-zzpk4"] Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124351 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-public-tls-certs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124496 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-config-data\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124542 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-combined-ca-bundle\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124566 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea51b4d4-7910-45a6-8320-a299877df017-logs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124607 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-config-data-custom\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124644 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdz4z\" (UniqueName: \"kubernetes.io/projected/ea51b4d4-7910-45a6-8320-a299877df017-kube-api-access-mdz4z\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.124717 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-internal-tls-certs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227591 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-internal-tls-certs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227690 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-public-tls-certs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227796 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-config-data\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227838 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-combined-ca-bundle\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227866 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea51b4d4-7910-45a6-8320-a299877df017-logs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227911 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-config-data-custom\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.227944 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdz4z\" (UniqueName: \"kubernetes.io/projected/ea51b4d4-7910-45a6-8320-a299877df017-kube-api-access-mdz4z\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.257080 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea51b4d4-7910-45a6-8320-a299877df017-logs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.261880 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-public-tls-certs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.262175 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-config-data-custom\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.266169 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-config-data\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.266913 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-combined-ca-bundle\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.284792 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdz4z\" (UniqueName: \"kubernetes.io/projected/ea51b4d4-7910-45a6-8320-a299877df017-kube-api-access-mdz4z\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.308841 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea51b4d4-7910-45a6-8320-a299877df017-internal-tls-certs\") pod \"barbican-api-54f666d7d6-zzpk4\" (UID: \"ea51b4d4-7910-45a6-8320-a299877df017\") " pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.587677 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:25 crc kubenswrapper[4710]: I1002 13:08:25.872469 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.047817 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-config-data\") pod \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.048164 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-combined-ca-bundle\") pod \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.048231 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrzkv\" (UniqueName: \"kubernetes.io/projected/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-kube-api-access-hrzkv\") pod \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.048326 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-etc-machine-id\") pod \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.048356 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-db-sync-config-data\") pod \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.048403 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-scripts\") pod \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\" (UID: \"6152e5af-f395-4bca-9aa1-19d17ec0ffa2\") " Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.049165 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6152e5af-f395-4bca-9aa1-19d17ec0ffa2" (UID: "6152e5af-f395-4bca-9aa1-19d17ec0ffa2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.058292 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6152e5af-f395-4bca-9aa1-19d17ec0ffa2" (UID: "6152e5af-f395-4bca-9aa1-19d17ec0ffa2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.058605 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-kube-api-access-hrzkv" (OuterVolumeSpecName: "kube-api-access-hrzkv") pod "6152e5af-f395-4bca-9aa1-19d17ec0ffa2" (UID: "6152e5af-f395-4bca-9aa1-19d17ec0ffa2"). InnerVolumeSpecName "kube-api-access-hrzkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.063935 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-scripts" (OuterVolumeSpecName: "scripts") pod "6152e5af-f395-4bca-9aa1-19d17ec0ffa2" (UID: "6152e5af-f395-4bca-9aa1-19d17ec0ffa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.124298 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6152e5af-f395-4bca-9aa1-19d17ec0ffa2" (UID: "6152e5af-f395-4bca-9aa1-19d17ec0ffa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.157553 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.157586 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrzkv\" (UniqueName: \"kubernetes.io/projected/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-kube-api-access-hrzkv\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.157599 4710 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.157611 4710 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.157623 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.312915 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-config-data" (OuterVolumeSpecName: "config-data") pod "6152e5af-f395-4bca-9aa1-19d17ec0ffa2" (UID: "6152e5af-f395-4bca-9aa1-19d17ec0ffa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.339443 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4m9cc" event={"ID":"6152e5af-f395-4bca-9aa1-19d17ec0ffa2","Type":"ContainerDied","Data":"3fe04575f67980db575ed297ed21e8ea7202bb48311456d9f337e4358e02f285"} Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.339492 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fe04575f67980db575ed297ed21e8ea7202bb48311456d9f337e4358e02f285" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.339573 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4m9cc" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.360510 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6152e5af-f395-4bca-9aa1-19d17ec0ffa2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:26 crc kubenswrapper[4710]: I1002 13:08:26.894331 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54f666d7d6-zzpk4"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.267824 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:27 crc kubenswrapper[4710]: E1002 13:08:27.268615 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" containerName="cinder-db-sync" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.268636 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" containerName="cinder-db-sync" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.269346 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" containerName="cinder-db-sync" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.270506 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.280613 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-c7scg" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.280922 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.281084 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.287150 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.302391 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.397682 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-scripts\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.397779 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.397855 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc9mk\" (UniqueName: \"kubernetes.io/projected/770412a4-b41f-4d11-986c-f585fae01d7c-kube-api-access-xc9mk\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.397932 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/770412a4-b41f-4d11-986c-f585fae01d7c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.397984 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.398033 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.423035 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8bffbc5c-kxntj"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.452098 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9bc865865-4f875"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.468186 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" event={"ID":"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd","Type":"ContainerStarted","Data":"312c892bb210f341de71d2b963ecb0b473c20e2acebe6c046daafeb2de3c8814"} Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.468253 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.468346 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.499497 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" event={"ID":"593ff4da-0e41-460e-9b1c-b8d0d032d492","Type":"ContainerStarted","Data":"6339db4602a5dfbb7bafc84149f31305e21c38e7e1cd5da68ca90147d24ac8aa"} Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.501243 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/770412a4-b41f-4d11-986c-f585fae01d7c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.501295 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.501333 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.501380 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-scripts\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.501401 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.501451 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc9mk\" (UniqueName: \"kubernetes.io/projected/770412a4-b41f-4d11-986c-f585fae01d7c-kube-api-access-xc9mk\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.505212 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/770412a4-b41f-4d11-986c-f585fae01d7c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.515974 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.516674 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9bc865865-4f875"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.531627 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" podStartSLOduration=6.531604282 podStartE2EDuration="6.531604282s" podCreationTimestamp="2025-10-02 13:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:27.49299737 +0000 UTC m=+1151.599408253" watchObservedRunningTime="2025-10-02 13:08:27.531604282 +0000 UTC m=+1151.638015165" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.549232 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.549646 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-scripts\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.550557 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerStarted","Data":"bf810ce4d53e2f76e3258d11df09631564767962abf0f8557e31cc29e316cf32"} Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.550879 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.552170 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.559298 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f666d7d6-zzpk4" event={"ID":"ea51b4d4-7910-45a6-8320-a299877df017","Type":"ContainerStarted","Data":"c7aba2f0f31da9a94430f63403c00b623ffa285c24ca8c15744ce3bf96aa0f0e"} Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.559482 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.563356 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc9mk\" (UniqueName: \"kubernetes.io/projected/770412a4-b41f-4d11-986c-f585fae01d7c-kube-api-access-xc9mk\") pod \"cinder-scheduler-0\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.564238 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.564367 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.599681 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f7dcd7d87-c6j6p" event={"ID":"ddcb91c7-f839-4756-9714-bf04aaf2382a","Type":"ContainerStarted","Data":"1026859b67de68f65d8eb85f1cb3ef3ac0796470894da2dba02f849a22ee57f9"} Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.608489 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.632894 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-svc\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.633170 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ldgp\" (UniqueName: \"kubernetes.io/projected/335e25f6-16e6-4aaa-a1b0-32595ab2591e-kube-api-access-8ldgp\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.633333 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-nb\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.633471 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-config\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.633571 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-swift-storage-0\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.633733 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-sb\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736281 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-svc\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736618 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02ac8fc8-b3db-4908-9322-ff4318f21a82-logs\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736641 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ldgp\" (UniqueName: \"kubernetes.io/projected/335e25f6-16e6-4aaa-a1b0-32595ab2591e-kube-api-access-8ldgp\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736667 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57mvw\" (UniqueName: \"kubernetes.io/projected/02ac8fc8-b3db-4908-9322-ff4318f21a82-kube-api-access-57mvw\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736692 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736721 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-nb\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736771 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data-custom\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736790 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-config\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736806 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736826 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ac8fc8-b3db-4908-9322-ff4318f21a82-etc-machine-id\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736842 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-swift-storage-0\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736877 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-scripts\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.736919 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-sb\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.737826 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-sb\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.738355 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-config\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.739069 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-svc\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.740106 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-swift-storage-0\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.743565 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-nb\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.777664 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ldgp\" (UniqueName: \"kubernetes.io/projected/335e25f6-16e6-4aaa-a1b0-32595ab2591e-kube-api-access-8ldgp\") pod \"dnsmasq-dns-9bc865865-4f875\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.838289 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02ac8fc8-b3db-4908-9322-ff4318f21a82-logs\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.847240 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57mvw\" (UniqueName: \"kubernetes.io/projected/02ac8fc8-b3db-4908-9322-ff4318f21a82-kube-api-access-57mvw\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.847310 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.847432 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data-custom\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.847471 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.847507 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ac8fc8-b3db-4908-9322-ff4318f21a82-etc-machine-id\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.847583 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-scripts\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.841163 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02ac8fc8-b3db-4908-9322-ff4318f21a82-logs\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.851481 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ac8fc8-b3db-4908-9322-ff4318f21a82-etc-machine-id\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.845043 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.854961 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.860576 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-scripts\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.867034 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.876337 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data-custom\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.899717 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57mvw\" (UniqueName: \"kubernetes.io/projected/02ac8fc8-b3db-4908-9322-ff4318f21a82-kube-api-access-57mvw\") pod \"cinder-api-0\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " pod="openstack/cinder-api-0" Oct 02 13:08:27 crc kubenswrapper[4710]: I1002 13:08:27.929284 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.030661 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.049637 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8cd8bc89b-fpjvs" Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.413156 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.570498 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9bc865865-4f875"] Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.624133 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f666d7d6-zzpk4" event={"ID":"ea51b4d4-7910-45a6-8320-a299877df017","Type":"ContainerStarted","Data":"9b39eb3e33faae19e965cc76efd614903913edfbc8eff1fadf1a2a1678c39f8e"} Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.626611 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f7dcd7d87-c6j6p" event={"ID":"ddcb91c7-f839-4756-9714-bf04aaf2382a","Type":"ContainerStarted","Data":"a65eb8361fd8104d89de35258db3efaff4cc13f3f21f50f1eef9a22dfdc40c96"} Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.651015 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerName="dnsmasq-dns" containerID="cri-o://312c892bb210f341de71d2b963ecb0b473c20e2acebe6c046daafeb2de3c8814" gracePeriod=10 Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.651853 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" event={"ID":"593ff4da-0e41-460e-9b1c-b8d0d032d492","Type":"ContainerStarted","Data":"6d010601ed065535439e169d7dee55a7e9cebb1de2415ad9ff6cfe95fedd8176"} Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.667546 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-f7dcd7d87-c6j6p" podStartSLOduration=3.90541865 podStartE2EDuration="7.667525423s" podCreationTimestamp="2025-10-02 13:08:21 +0000 UTC" firstStartedPulling="2025-10-02 13:08:22.64239832 +0000 UTC m=+1146.748809203" lastFinishedPulling="2025-10-02 13:08:26.404505093 +0000 UTC m=+1150.510915976" observedRunningTime="2025-10-02 13:08:28.657173793 +0000 UTC m=+1152.763584696" watchObservedRunningTime="2025-10-02 13:08:28.667525423 +0000 UTC m=+1152.773936306" Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.708995 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5c86c47bf8-5k59q" podStartSLOduration=3.739699329 podStartE2EDuration="7.708970226s" podCreationTimestamp="2025-10-02 13:08:21 +0000 UTC" firstStartedPulling="2025-10-02 13:08:22.433689117 +0000 UTC m=+1146.540099990" lastFinishedPulling="2025-10-02 13:08:26.402960004 +0000 UTC m=+1150.509370887" observedRunningTime="2025-10-02 13:08:28.690496381 +0000 UTC m=+1152.796907264" watchObservedRunningTime="2025-10-02 13:08:28.708970226 +0000 UTC m=+1152.815381109" Oct 02 13:08:28 crc kubenswrapper[4710]: I1002 13:08:28.741344 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.222250 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.676682 4710 generic.go:334] "Generic (PLEG): container finished" podID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerID="312c892bb210f341de71d2b963ecb0b473c20e2acebe6c046daafeb2de3c8814" exitCode=0 Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.676999 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" event={"ID":"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd","Type":"ContainerDied","Data":"312c892bb210f341de71d2b963ecb0b473c20e2acebe6c046daafeb2de3c8814"} Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.687299 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"02ac8fc8-b3db-4908-9322-ff4318f21a82","Type":"ContainerStarted","Data":"9e68648c3fae9a0324664bc73173a1332f93071cdf6d97e5dee19b42f0f73657"} Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.717031 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"770412a4-b41f-4d11-986c-f585fae01d7c","Type":"ContainerStarted","Data":"ff9816cfc65a2a4aa625b3f6d446ab7e755382bd0b3fd41372444230f00465f1"} Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.805095 4710 generic.go:334] "Generic (PLEG): container finished" podID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerID="b5c52f2790e18500ac8222c04e234a2e4b874932f58a260273793e7b95afe6ef" exitCode=0 Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.805209 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bc865865-4f875" event={"ID":"335e25f6-16e6-4aaa-a1b0-32595ab2591e","Type":"ContainerDied","Data":"b5c52f2790e18500ac8222c04e234a2e4b874932f58a260273793e7b95afe6ef"} Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.805242 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bc865865-4f875" event={"ID":"335e25f6-16e6-4aaa-a1b0-32595ab2591e","Type":"ContainerStarted","Data":"d8d3dbd93dace3af58024d5126b46a224e9a5e8c8c85a6561bd9819cf748d4ec"} Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.839183 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f666d7d6-zzpk4" event={"ID":"ea51b4d4-7910-45a6-8320-a299877df017","Type":"ContainerStarted","Data":"33dcb3d16085177daca531e875b709030b5bacdec5e2db014a9b340ff51b6909"} Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.839267 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.857730 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:29 crc kubenswrapper[4710]: I1002 13:08:29.957828 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54f666d7d6-zzpk4" podStartSLOduration=5.95780739 podStartE2EDuration="5.95780739s" podCreationTimestamp="2025-10-02 13:08:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:29.937870938 +0000 UTC m=+1154.044281831" watchObservedRunningTime="2025-10-02 13:08:29.95780739 +0000 UTC m=+1154.064218273" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.490240 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.549679 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-svc\") pod \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.550096 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpkrl\" (UniqueName: \"kubernetes.io/projected/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-kube-api-access-wpkrl\") pod \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.550207 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-config\") pod \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.550236 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-sb\") pod \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.550312 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-swift-storage-0\") pod \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.550421 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-nb\") pod \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\" (UID: \"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd\") " Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.594516 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-kube-api-access-wpkrl" (OuterVolumeSpecName: "kube-api-access-wpkrl") pod "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" (UID: "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd"). InnerVolumeSpecName "kube-api-access-wpkrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.654313 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpkrl\" (UniqueName: \"kubernetes.io/projected/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-kube-api-access-wpkrl\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.775041 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-config" (OuterVolumeSpecName: "config") pod "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" (UID: "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.839965 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.840444 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8bffbc5c-kxntj" event={"ID":"410d2ef2-2ec1-4104-b6f1-b84b9a831dcd","Type":"ContainerDied","Data":"8abecb6ed1e69c5cbedeaee15282ef35b18954fabb442fda22d92be7861ffb96"} Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.840480 4710 scope.go:117] "RemoveContainer" containerID="312c892bb210f341de71d2b963ecb0b473c20e2acebe6c046daafeb2de3c8814" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.867861 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.917638 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" (UID: "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.969971 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:30 crc kubenswrapper[4710]: I1002 13:08:30.986998 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" (UID: "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.006583 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" (UID: "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.014092 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" (UID: "410d2ef2-2ec1-4104-b6f1-b84b9a831dcd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.072359 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.072392 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.072402 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.118456 4710 scope.go:117] "RemoveContainer" containerID="63d431214dd89183d41cf9cbf34267645fbd1fd0706bf82bb600375ca2c3512e" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.136661 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.164:8443: connect: connection refused" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.192757 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8bffbc5c-kxntj"] Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.220591 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c8bffbc5c-kxntj"] Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.248091 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.456213 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c89fc49-c8w65" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.524576 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f5fcfd868-8dmb9"] Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.526539 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7fc95bbdbd-gxqdq" Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.527901 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-f5fcfd868-8dmb9" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-api" containerID="cri-o://30bbc857c2287e3dbf3e081f710f3c5e5f86bed6b01d689f0e337fbd7f3d1f1c" gracePeriod=30 Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.528519 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-f5fcfd868-8dmb9" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-httpd" containerID="cri-o://ad530a2586f341efcb8a07c56a3915e8efc95ca122fbbd798351b9a66fa58c45" gracePeriod=30 Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.894076 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"02ac8fc8-b3db-4908-9322-ff4318f21a82","Type":"ContainerStarted","Data":"b3d79be00a538bc5e49dcc2998708519fe59a6073e4e64331ed8a88afaa78853"} Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.951059 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerStarted","Data":"73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7"} Oct 02 13:08:31 crc kubenswrapper[4710]: I1002 13:08:31.951372 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:08:32 crc kubenswrapper[4710]: I1002 13:08:32.012450 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bc865865-4f875" event={"ID":"335e25f6-16e6-4aaa-a1b0-32595ab2591e","Type":"ContainerStarted","Data":"d599285d2de11783bf93fa6dc9c6be10e9a0fe21db62e996d8ec62843903b065"} Oct 02 13:08:32 crc kubenswrapper[4710]: I1002 13:08:32.014030 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:32 crc kubenswrapper[4710]: I1002 13:08:32.019330 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.096867068 podStartE2EDuration="13.019303418s" podCreationTimestamp="2025-10-02 13:08:19 +0000 UTC" firstStartedPulling="2025-10-02 13:08:20.028790395 +0000 UTC m=+1144.135201268" lastFinishedPulling="2025-10-02 13:08:29.951226725 +0000 UTC m=+1154.057637618" observedRunningTime="2025-10-02 13:08:31.986872922 +0000 UTC m=+1156.093283805" watchObservedRunningTime="2025-10-02 13:08:32.019303418 +0000 UTC m=+1156.125714311" Oct 02 13:08:32 crc kubenswrapper[4710]: I1002 13:08:32.094956 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9bc865865-4f875" podStartSLOduration=5.094931422 podStartE2EDuration="5.094931422s" podCreationTimestamp="2025-10-02 13:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:32.07619635 +0000 UTC m=+1156.182607233" watchObservedRunningTime="2025-10-02 13:08:32.094931422 +0000 UTC m=+1156.201342305" Oct 02 13:08:32 crc kubenswrapper[4710]: I1002 13:08:32.926117 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" path="/var/lib/kubelet/pods/410d2ef2-2ec1-4104-b6f1-b84b9a831dcd/volumes" Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.073034 4710 generic.go:334] "Generic (PLEG): container finished" podID="e2022f45-9ec9-483a-bce6-85945738da7f" containerID="ad530a2586f341efcb8a07c56a3915e8efc95ca122fbbd798351b9a66fa58c45" exitCode=0 Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.073127 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f5fcfd868-8dmb9" event={"ID":"e2022f45-9ec9-483a-bce6-85945738da7f","Type":"ContainerDied","Data":"ad530a2586f341efcb8a07c56a3915e8efc95ca122fbbd798351b9a66fa58c45"} Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.107056 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api-log" containerID="cri-o://b3d79be00a538bc5e49dcc2998708519fe59a6073e4e64331ed8a88afaa78853" gracePeriod=30 Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.107433 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"02ac8fc8-b3db-4908-9322-ff4318f21a82","Type":"ContainerStarted","Data":"1c1314922533bec141ed8d13d3af2eedc05cae2eadb250a7f7776c45d549b771"} Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.108062 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.108517 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api" containerID="cri-o://1c1314922533bec141ed8d13d3af2eedc05cae2eadb250a7f7776c45d549b771" gracePeriod=30 Oct 02 13:08:33 crc kubenswrapper[4710]: I1002 13:08:33.134299 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.134279342 podStartE2EDuration="6.134279342s" podCreationTimestamp="2025-10-02 13:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:33.127309487 +0000 UTC m=+1157.233720370" watchObservedRunningTime="2025-10-02 13:08:33.134279342 +0000 UTC m=+1157.240690235" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.147366 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"02ac8fc8-b3db-4908-9322-ff4318f21a82","Type":"ContainerDied","Data":"b3d79be00a538bc5e49dcc2998708519fe59a6073e4e64331ed8a88afaa78853"} Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.147409 4710 generic.go:334] "Generic (PLEG): container finished" podID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerID="b3d79be00a538bc5e49dcc2998708519fe59a6073e4e64331ed8a88afaa78853" exitCode=143 Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.151412 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"770412a4-b41f-4d11-986c-f585fae01d7c","Type":"ContainerStarted","Data":"71633b08c3b5aae4bac4e04d2ea2ad8365b518d22b20aa87db560d45369cfb09"} Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.237477 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.422989 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.607804 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 13:08:34 crc kubenswrapper[4710]: E1002 13:08:34.608689 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerName="init" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.608713 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerName="init" Oct 02 13:08:34 crc kubenswrapper[4710]: E1002 13:08:34.608767 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerName="dnsmasq-dns" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.608776 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerName="dnsmasq-dns" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.608978 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="410d2ef2-2ec1-4104-b6f1-b84b9a831dcd" containerName="dnsmasq-dns" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.609817 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.613904 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-tz95t" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.621455 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.622031 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.631993 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.790979 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71a2a3d5-5a6a-4f31-8df3-f087529d550d-openstack-config-secret\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.791068 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrc4v\" (UniqueName: \"kubernetes.io/projected/71a2a3d5-5a6a-4f31-8df3-f087529d550d-kube-api-access-jrc4v\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.791093 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a2a3d5-5a6a-4f31-8df3-f087529d550d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.791114 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71a2a3d5-5a6a-4f31-8df3-f087529d550d-openstack-config\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.843148 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.893000 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71a2a3d5-5a6a-4f31-8df3-f087529d550d-openstack-config-secret\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.893179 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrc4v\" (UniqueName: \"kubernetes.io/projected/71a2a3d5-5a6a-4f31-8df3-f087529d550d-kube-api-access-jrc4v\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.893219 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a2a3d5-5a6a-4f31-8df3-f087529d550d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.893243 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71a2a3d5-5a6a-4f31-8df3-f087529d550d-openstack-config\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.895894 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71a2a3d5-5a6a-4f31-8df3-f087529d550d-openstack-config\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.901332 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71a2a3d5-5a6a-4f31-8df3-f087529d550d-openstack-config-secret\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.918272 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a2a3d5-5a6a-4f31-8df3-f087529d550d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.935308 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrc4v\" (UniqueName: \"kubernetes.io/projected/71a2a3d5-5a6a-4f31-8df3-f087529d550d-kube-api-access-jrc4v\") pod \"openstackclient\" (UID: \"71a2a3d5-5a6a-4f31-8df3-f087529d550d\") " pod="openstack/openstackclient" Oct 02 13:08:34 crc kubenswrapper[4710]: I1002 13:08:34.970191 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 13:08:35 crc kubenswrapper[4710]: I1002 13:08:35.615439 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 13:08:35 crc kubenswrapper[4710]: I1002 13:08:35.906074 4710 scope.go:117] "RemoveContainer" containerID="341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec" Oct 02 13:08:35 crc kubenswrapper[4710]: E1002 13:08:35.906310 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:36 crc kubenswrapper[4710]: I1002 13:08:36.216726 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"770412a4-b41f-4d11-986c-f585fae01d7c","Type":"ContainerStarted","Data":"3ef4f270e7b999b7af90be362bdfaff814a78345b72e8869d4bdf342f9ce8966"} Oct 02 13:08:36 crc kubenswrapper[4710]: I1002 13:08:36.223530 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"71a2a3d5-5a6a-4f31-8df3-f087529d550d","Type":"ContainerStarted","Data":"6c721355299ef6f4426d83aa4dd378c4849132f3144201a753be1febd6dbaa71"} Oct 02 13:08:36 crc kubenswrapper[4710]: I1002 13:08:36.938162 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.351934361 podStartE2EDuration="9.938145247s" podCreationTimestamp="2025-10-02 13:08:27 +0000 UTC" firstStartedPulling="2025-10-02 13:08:28.825248853 +0000 UTC m=+1152.931659736" lastFinishedPulling="2025-10-02 13:08:32.411459739 +0000 UTC m=+1156.517870622" observedRunningTime="2025-10-02 13:08:36.247418191 +0000 UTC m=+1160.353829074" watchObservedRunningTime="2025-10-02 13:08:36.938145247 +0000 UTC m=+1161.044556130" Oct 02 13:08:37 crc kubenswrapper[4710]: I1002 13:08:37.610976 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 13:08:37 crc kubenswrapper[4710]: I1002 13:08:37.845936 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:08:37 crc kubenswrapper[4710]: I1002 13:08:37.917962 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b4d46b6c-l4bbj"] Oct 02 13:08:37 crc kubenswrapper[4710]: I1002 13:08:37.918244 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerName="dnsmasq-dns" containerID="cri-o://276ed2623050743a03d0a90e28cb257b01449d397ca1836d8531fbe9afd03248" gracePeriod=10 Oct 02 13:08:37 crc kubenswrapper[4710]: I1002 13:08:37.933481 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54f666d7d6-zzpk4" Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.007675 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-844bdb96b6-wthf8"] Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.007970 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-844bdb96b6-wthf8" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api-log" containerID="cri-o://5fd49cd51631f5f283cd0ad640e6fe5d344ff2ea3219e3b57ecb8c2cc2517c4b" gracePeriod=30 Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.008434 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-844bdb96b6-wthf8" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api" containerID="cri-o://28bcdefa12b42f75cfb4313625fc0a6c9ee02596c8f22e5d194091d3ff8d682f" gracePeriod=30 Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.263241 4710 generic.go:334] "Generic (PLEG): container finished" podID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerID="5fd49cd51631f5f283cd0ad640e6fe5d344ff2ea3219e3b57ecb8c2cc2517c4b" exitCode=143 Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.263309 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-844bdb96b6-wthf8" event={"ID":"66d6931f-1351-4285-a8fc-e1280f21dee6","Type":"ContainerDied","Data":"5fd49cd51631f5f283cd0ad640e6fe5d344ff2ea3219e3b57ecb8c2cc2517c4b"} Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.302006 4710 generic.go:334] "Generic (PLEG): container finished" podID="e2022f45-9ec9-483a-bce6-85945738da7f" containerID="30bbc857c2287e3dbf3e081f710f3c5e5f86bed6b01d689f0e337fbd7f3d1f1c" exitCode=0 Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.302100 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f5fcfd868-8dmb9" event={"ID":"e2022f45-9ec9-483a-bce6-85945738da7f","Type":"ContainerDied","Data":"30bbc857c2287e3dbf3e081f710f3c5e5f86bed6b01d689f0e337fbd7f3d1f1c"} Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.328648 4710 generic.go:334] "Generic (PLEG): container finished" podID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerID="276ed2623050743a03d0a90e28cb257b01449d397ca1836d8531fbe9afd03248" exitCode=0 Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.329498 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" event={"ID":"48923995-d4ba-4562-8db2-26d5d34a0be5","Type":"ContainerDied","Data":"276ed2623050743a03d0a90e28cb257b01449d397ca1836d8531fbe9afd03248"} Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.828716 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.887254 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.911805 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-config\") pod \"e2022f45-9ec9-483a-bce6-85945738da7f\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.911990 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-httpd-config\") pod \"e2022f45-9ec9-483a-bce6-85945738da7f\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.912177 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vwnh\" (UniqueName: \"kubernetes.io/projected/e2022f45-9ec9-483a-bce6-85945738da7f-kube-api-access-4vwnh\") pod \"e2022f45-9ec9-483a-bce6-85945738da7f\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.912263 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-ovndb-tls-certs\") pod \"e2022f45-9ec9-483a-bce6-85945738da7f\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.912427 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-combined-ca-bundle\") pod \"e2022f45-9ec9-483a-bce6-85945738da7f\" (UID: \"e2022f45-9ec9-483a-bce6-85945738da7f\") " Oct 02 13:08:38 crc kubenswrapper[4710]: I1002 13:08:38.929695 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e2022f45-9ec9-483a-bce6-85945738da7f" (UID: "e2022f45-9ec9-483a-bce6-85945738da7f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:38.990989 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2022f45-9ec9-483a-bce6-85945738da7f-kube-api-access-4vwnh" (OuterVolumeSpecName: "kube-api-access-4vwnh") pod "e2022f45-9ec9-483a-bce6-85945738da7f" (UID: "e2022f45-9ec9-483a-bce6-85945738da7f"). InnerVolumeSpecName "kube-api-access-4vwnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.026433 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-config" (OuterVolumeSpecName: "config") pod "e2022f45-9ec9-483a-bce6-85945738da7f" (UID: "e2022f45-9ec9-483a-bce6-85945738da7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.031273 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhhmn\" (UniqueName: \"kubernetes.io/projected/48923995-d4ba-4562-8db2-26d5d34a0be5-kube-api-access-rhhmn\") pod \"48923995-d4ba-4562-8db2-26d5d34a0be5\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.031406 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-config\") pod \"48923995-d4ba-4562-8db2-26d5d34a0be5\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.031428 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-nb\") pod \"48923995-d4ba-4562-8db2-26d5d34a0be5\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.031456 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-svc\") pod \"48923995-d4ba-4562-8db2-26d5d34a0be5\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.031549 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-swift-storage-0\") pod \"48923995-d4ba-4562-8db2-26d5d34a0be5\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.031609 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-sb\") pod \"48923995-d4ba-4562-8db2-26d5d34a0be5\" (UID: \"48923995-d4ba-4562-8db2-26d5d34a0be5\") " Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.032672 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.032693 4710 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.032706 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vwnh\" (UniqueName: \"kubernetes.io/projected/e2022f45-9ec9-483a-bce6-85945738da7f-kube-api-access-4vwnh\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.107276 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48923995-d4ba-4562-8db2-26d5d34a0be5-kube-api-access-rhhmn" (OuterVolumeSpecName: "kube-api-access-rhhmn") pod "48923995-d4ba-4562-8db2-26d5d34a0be5" (UID: "48923995-d4ba-4562-8db2-26d5d34a0be5"). InnerVolumeSpecName "kube-api-access-rhhmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.136936 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhhmn\" (UniqueName: \"kubernetes.io/projected/48923995-d4ba-4562-8db2-26d5d34a0be5-kube-api-access-rhhmn\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.161902 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2022f45-9ec9-483a-bce6-85945738da7f" (UID: "e2022f45-9ec9-483a-bce6-85945738da7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.242796 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.265098 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "48923995-d4ba-4562-8db2-26d5d34a0be5" (UID: "48923995-d4ba-4562-8db2-26d5d34a0be5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.278301 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48923995-d4ba-4562-8db2-26d5d34a0be5" (UID: "48923995-d4ba-4562-8db2-26d5d34a0be5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.305836 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "48923995-d4ba-4562-8db2-26d5d34a0be5" (UID: "48923995-d4ba-4562-8db2-26d5d34a0be5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.315676 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e2022f45-9ec9-483a-bce6-85945738da7f" (UID: "e2022f45-9ec9-483a-bce6-85945738da7f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.344346 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.344383 4710 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2022f45-9ec9-483a-bce6-85945738da7f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.344397 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.344410 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.354383 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48923995-d4ba-4562-8db2-26d5d34a0be5" (UID: "48923995-d4ba-4562-8db2-26d5d34a0be5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.360349 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f5fcfd868-8dmb9" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.361202 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f5fcfd868-8dmb9" event={"ID":"e2022f45-9ec9-483a-bce6-85945738da7f","Type":"ContainerDied","Data":"6f1aa040b5ac865f726204ed33a938f1af8a03bf1ae676bc109f26d14d9b9df7"} Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.361267 4710 scope.go:117] "RemoveContainer" containerID="ad530a2586f341efcb8a07c56a3915e8efc95ca122fbbd798351b9a66fa58c45" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.368479 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" event={"ID":"48923995-d4ba-4562-8db2-26d5d34a0be5","Type":"ContainerDied","Data":"94a1f583a0a32c8a1aec8fe50b8aecae0281a77bba4dec550e5ebd50a5b68360"} Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.368609 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b4d46b6c-l4bbj" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.368902 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-config" (OuterVolumeSpecName: "config") pod "48923995-d4ba-4562-8db2-26d5d34a0be5" (UID: "48923995-d4ba-4562-8db2-26d5d34a0be5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.401530 4710 scope.go:117] "RemoveContainer" containerID="30bbc857c2287e3dbf3e081f710f3c5e5f86bed6b01d689f0e337fbd7f3d1f1c" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.402889 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f5fcfd868-8dmb9"] Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.415560 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f5fcfd868-8dmb9"] Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.424849 4710 scope.go:117] "RemoveContainer" containerID="276ed2623050743a03d0a90e28cb257b01449d397ca1836d8531fbe9afd03248" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.445779 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.445811 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48923995-d4ba-4562-8db2-26d5d34a0be5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.461515 4710 scope.go:117] "RemoveContainer" containerID="4e136af53a5996f4f900357b870ceca06f7437fbf36fb4667abd6c76f1158937" Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.699542 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b4d46b6c-l4bbj"] Oct 02 13:08:39 crc kubenswrapper[4710]: I1002 13:08:39.707475 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59b4d46b6c-l4bbj"] Oct 02 13:08:40 crc kubenswrapper[4710]: I1002 13:08:40.914462 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" path="/var/lib/kubelet/pods/48923995-d4ba-4562-8db2-26d5d34a0be5/volumes" Oct 02 13:08:40 crc kubenswrapper[4710]: I1002 13:08:40.915339 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" path="/var/lib/kubelet/pods/e2022f45-9ec9-483a-bce6-85945738da7f/volumes" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.136339 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.164:8443: connect: connection refused" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.392639 4710 generic.go:334] "Generic (PLEG): container finished" podID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerID="28bcdefa12b42f75cfb4313625fc0a6c9ee02596c8f22e5d194091d3ff8d682f" exitCode=0 Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.392824 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-844bdb96b6-wthf8" event={"ID":"66d6931f-1351-4285-a8fc-e1280f21dee6","Type":"ContainerDied","Data":"28bcdefa12b42f75cfb4313625fc0a6c9ee02596c8f22e5d194091d3ff8d682f"} Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.600259 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.797088 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.893824 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-combined-ca-bundle\") pod \"66d6931f-1351-4285-a8fc-e1280f21dee6\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.895049 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data-custom\") pod \"66d6931f-1351-4285-a8fc-e1280f21dee6\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.897084 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6931f-1351-4285-a8fc-e1280f21dee6-logs\") pod \"66d6931f-1351-4285-a8fc-e1280f21dee6\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.897284 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7g5h\" (UniqueName: \"kubernetes.io/projected/66d6931f-1351-4285-a8fc-e1280f21dee6-kube-api-access-h7g5h\") pod \"66d6931f-1351-4285-a8fc-e1280f21dee6\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.897417 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data\") pod \"66d6931f-1351-4285-a8fc-e1280f21dee6\" (UID: \"66d6931f-1351-4285-a8fc-e1280f21dee6\") " Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.897586 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d6931f-1351-4285-a8fc-e1280f21dee6-logs" (OuterVolumeSpecName: "logs") pod "66d6931f-1351-4285-a8fc-e1280f21dee6" (UID: "66d6931f-1351-4285-a8fc-e1280f21dee6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.898377 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6931f-1351-4285-a8fc-e1280f21dee6-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.900947 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d6931f-1351-4285-a8fc-e1280f21dee6-kube-api-access-h7g5h" (OuterVolumeSpecName: "kube-api-access-h7g5h") pod "66d6931f-1351-4285-a8fc-e1280f21dee6" (UID: "66d6931f-1351-4285-a8fc-e1280f21dee6"). InnerVolumeSpecName "kube-api-access-h7g5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.901545 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "66d6931f-1351-4285-a8fc-e1280f21dee6" (UID: "66d6931f-1351-4285-a8fc-e1280f21dee6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.934505 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66d6931f-1351-4285-a8fc-e1280f21dee6" (UID: "66d6931f-1351-4285-a8fc-e1280f21dee6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:41 crc kubenswrapper[4710]: I1002 13:08:41.975718 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data" (OuterVolumeSpecName: "config-data") pod "66d6931f-1351-4285-a8fc-e1280f21dee6" (UID: "66d6931f-1351-4285-a8fc-e1280f21dee6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.000336 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7g5h\" (UniqueName: \"kubernetes.io/projected/66d6931f-1351-4285-a8fc-e1280f21dee6-kube-api-access-h7g5h\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.000372 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.000382 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.000390 4710 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66d6931f-1351-4285-a8fc-e1280f21dee6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.092739 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.093551 4710 scope.go:117] "RemoveContainer" containerID="341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.093769 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.404318 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-844bdb96b6-wthf8" event={"ID":"66d6931f-1351-4285-a8fc-e1280f21dee6","Type":"ContainerDied","Data":"930bb7269bce50194d6b64f6d3c2da70c70872110c60a9469e2d65b982dcb152"} Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.404368 4710 scope.go:117] "RemoveContainer" containerID="28bcdefa12b42f75cfb4313625fc0a6c9ee02596c8f22e5d194091d3ff8d682f" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.404464 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-844bdb96b6-wthf8" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.436097 4710 scope.go:117] "RemoveContainer" containerID="5fd49cd51631f5f283cd0ad640e6fe5d344ff2ea3219e3b57ecb8c2cc2517c4b" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.438191 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-844bdb96b6-wthf8"] Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.472268 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-844bdb96b6-wthf8"] Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.755535 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.807548 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:42 crc kubenswrapper[4710]: I1002 13:08:42.937049 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" path="/var/lib/kubelet/pods/66d6931f-1351-4285-a8fc-e1280f21dee6/volumes" Oct 02 13:08:43 crc kubenswrapper[4710]: I1002 13:08:43.421150 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerStarted","Data":"8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44"} Oct 02 13:08:43 crc kubenswrapper[4710]: I1002 13:08:43.421998 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="cinder-scheduler" containerID="cri-o://71633b08c3b5aae4bac4e04d2ea2ad8365b518d22b20aa87db560d45369cfb09" gracePeriod=30 Oct 02 13:08:43 crc kubenswrapper[4710]: I1002 13:08:43.422109 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="probe" containerID="cri-o://3ef4f270e7b999b7af90be362bdfaff814a78345b72e8869d4bdf342f9ce8966" gracePeriod=30 Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.449683 4710 generic.go:334] "Generic (PLEG): container finished" podID="770412a4-b41f-4d11-986c-f585fae01d7c" containerID="3ef4f270e7b999b7af90be362bdfaff814a78345b72e8869d4bdf342f9ce8966" exitCode=0 Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.449786 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"770412a4-b41f-4d11-986c-f585fae01d7c","Type":"ContainerDied","Data":"3ef4f270e7b999b7af90be362bdfaff814a78345b72e8869d4bdf342f9ce8966"} Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453113 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-76f964d7cf-dhr2b"] Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.453596 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-api" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453616 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-api" Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.453632 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-httpd" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453640 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-httpd" Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.453655 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453661 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api" Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.453678 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api-log" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453684 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api-log" Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.453694 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerName="dnsmasq-dns" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453699 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerName="dnsmasq-dns" Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.453713 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerName="init" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453719 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerName="init" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453920 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api-log" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453936 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="48923995-d4ba-4562-8db2-26d5d34a0be5" containerName="dnsmasq-dns" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453948 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d6931f-1351-4285-a8fc-e1280f21dee6" containerName="barbican-api" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453961 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-api" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.453972 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2022f45-9ec9-483a-bce6-85945738da7f" containerName="neutron-httpd" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.456976 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.461725 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.462437 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.462658 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.471902 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-76f964d7cf-dhr2b"] Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.553208 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-combined-ca-bundle\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.553269 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-internal-tls-certs\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.553730 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-config-data\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.553886 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwc4h\" (UniqueName: \"kubernetes.io/projected/407d1688-f160-4ee8-934d-9c76f353e4b9-kube-api-access-vwc4h\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.554122 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/407d1688-f160-4ee8-934d-9c76f353e4b9-log-httpd\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.554211 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-public-tls-certs\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.554290 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/407d1688-f160-4ee8-934d-9c76f353e4b9-etc-swift\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.554364 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/407d1688-f160-4ee8-934d-9c76f353e4b9-run-httpd\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655619 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-config-data\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655692 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwc4h\" (UniqueName: \"kubernetes.io/projected/407d1688-f160-4ee8-934d-9c76f353e4b9-kube-api-access-vwc4h\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655774 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/407d1688-f160-4ee8-934d-9c76f353e4b9-log-httpd\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655795 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-public-tls-certs\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655812 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/407d1688-f160-4ee8-934d-9c76f353e4b9-etc-swift\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655828 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/407d1688-f160-4ee8-934d-9c76f353e4b9-run-httpd\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655872 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-combined-ca-bundle\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.655911 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-internal-tls-certs\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.656647 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/407d1688-f160-4ee8-934d-9c76f353e4b9-run-httpd\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.656845 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/407d1688-f160-4ee8-934d-9c76f353e4b9-log-httpd\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.661633 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-public-tls-certs\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.662101 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/407d1688-f160-4ee8-934d-9c76f353e4b9-etc-swift\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.662923 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-config-data\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.664371 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-internal-tls-certs\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.673564 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwc4h\" (UniqueName: \"kubernetes.io/projected/407d1688-f160-4ee8-934d-9c76f353e4b9-kube-api-access-vwc4h\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.679605 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/407d1688-f160-4ee8-934d-9c76f353e4b9-combined-ca-bundle\") pod \"swift-proxy-76f964d7cf-dhr2b\" (UID: \"407d1688-f160-4ee8-934d-9c76f353e4b9\") " pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: I1002 13:08:44.812422 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:44 crc kubenswrapper[4710]: E1002 13:08:44.912654 4710 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: , extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_aed5f89f-6790-4f1f-9ed2-2f075f8d4647/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_aed5f89f-6790-4f1f-9ed2-2f075f8d4647/ceilometer-notification-agent/0.log: no such file or directory Oct 02 13:08:45 crc kubenswrapper[4710]: I1002 13:08:45.464811 4710 generic.go:334] "Generic (PLEG): container finished" podID="770412a4-b41f-4d11-986c-f585fae01d7c" containerID="71633b08c3b5aae4bac4e04d2ea2ad8365b518d22b20aa87db560d45369cfb09" exitCode=0 Oct 02 13:08:45 crc kubenswrapper[4710]: I1002 13:08:45.464863 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"770412a4-b41f-4d11-986c-f585fae01d7c","Type":"ContainerDied","Data":"71633b08c3b5aae4bac4e04d2ea2ad8365b518d22b20aa87db560d45369cfb09"} Oct 02 13:08:45 crc kubenswrapper[4710]: W1002 13:08:45.768125 4710 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod410d2ef2_2ec1_4104_b6f1_b84b9a831dcd.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod410d2ef2_2ec1_4104_b6f1_b84b9a831dcd.slice: no such file or directory Oct 02 13:08:45 crc kubenswrapper[4710]: W1002 13:08:45.768183 4710 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d6931f_1351_4285_a8fc_e1280f21dee6.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d6931f_1351_4285_a8fc_e1280f21dee6.slice: no such file or directory Oct 02 13:08:46 crc kubenswrapper[4710]: I1002 13:08:46.481820 4710 generic.go:334] "Generic (PLEG): container finished" podID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerID="a42ca0c2045d5f637fdd39fdecb14ac1b0dd3e6fe65cc97a6a353834d22c6f67" exitCode=137 Oct 02 13:08:46 crc kubenswrapper[4710]: I1002 13:08:46.482227 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f85867556-bdslh" event={"ID":"00020afa-1e01-4a2c-924e-2aa55bac0e85","Type":"ContainerDied","Data":"a42ca0c2045d5f637fdd39fdecb14ac1b0dd3e6fe65cc97a6a353834d22c6f67"} Oct 02 13:08:47 crc kubenswrapper[4710]: I1002 13:08:47.473502 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:47 crc kubenswrapper[4710]: I1002 13:08:47.474181 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-central-agent" containerID="cri-o://b4d999f5769fbff464e6d35ac58c689edf5e54b43e48465dea610111f633ef4e" gracePeriod=30 Oct 02 13:08:47 crc kubenswrapper[4710]: I1002 13:08:47.475021 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="proxy-httpd" containerID="cri-o://73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7" gracePeriod=30 Oct 02 13:08:47 crc kubenswrapper[4710]: I1002 13:08:47.475093 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="sg-core" containerID="cri-o://bf810ce4d53e2f76e3258d11df09631564767962abf0f8557e31cc29e316cf32" gracePeriod=30 Oct 02 13:08:47 crc kubenswrapper[4710]: I1002 13:08:47.475140 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-notification-agent" containerID="cri-o://a7572dc21b0e3541f4fd4ee817688f46e16ed0b9c70d512f30df4b8982ca8772" gracePeriod=30 Oct 02 13:08:47 crc kubenswrapper[4710]: I1002 13:08:47.481970 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.532176 4710 generic.go:334] "Generic (PLEG): container finished" podID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" exitCode=1 Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.532544 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerDied","Data":"8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44"} Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.532586 4710 scope.go:117] "RemoveContainer" containerID="341a2696153fa568f4dde1379003113b77619da95c247541547071e24cc6d6ec" Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.533333 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:08:48 crc kubenswrapper[4710]: E1002 13:08:48.533625 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.540412 4710 generic.go:334] "Generic (PLEG): container finished" podID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerID="73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7" exitCode=0 Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.540451 4710 generic.go:334] "Generic (PLEG): container finished" podID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerID="bf810ce4d53e2f76e3258d11df09631564767962abf0f8557e31cc29e316cf32" exitCode=2 Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.540462 4710 generic.go:334] "Generic (PLEG): container finished" podID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerID="b4d999f5769fbff464e6d35ac58c689edf5e54b43e48465dea610111f633ef4e" exitCode=0 Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.540486 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerDied","Data":"73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7"} Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.540514 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerDied","Data":"bf810ce4d53e2f76e3258d11df09631564767962abf0f8557e31cc29e316cf32"} Oct 02 13:08:48 crc kubenswrapper[4710]: I1002 13:08:48.540526 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerDied","Data":"b4d999f5769fbff464e6d35ac58c689edf5e54b43e48465dea610111f633ef4e"} Oct 02 13:08:49 crc kubenswrapper[4710]: I1002 13:08:49.555659 4710 generic.go:334] "Generic (PLEG): container finished" podID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerID="a7572dc21b0e3541f4fd4ee817688f46e16ed0b9c70d512f30df4b8982ca8772" exitCode=0 Oct 02 13:08:49 crc kubenswrapper[4710]: I1002 13:08:49.555698 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerDied","Data":"a7572dc21b0e3541f4fd4ee817688f46e16ed0b9c70d512f30df4b8982ca8772"} Oct 02 13:08:49 crc kubenswrapper[4710]: I1002 13:08:49.583123 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.177:3000/\": dial tcp 10.217.0.177:3000: connect: connection refused" Oct 02 13:08:51 crc kubenswrapper[4710]: I1002 13:08:51.136549 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5f85867556-bdslh" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.164:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.164:8443: connect: connection refused" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.089333 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-76f964d7cf-dhr2b"] Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.100094 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.100147 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.100990 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:08:52 crc kubenswrapper[4710]: E1002 13:08:52.101321 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.185883 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.186198 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="27670d96-0e34-41ae-a54c-2bb7f19b1c6e" containerName="kube-state-metrics" containerID="cri-o://35739b19eacea9fcf2439673fc3c3a190f8a8a108b09dc5fe1742feecb333446" gracePeriod=30 Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.530259 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.530575 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.626888 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-76f964d7cf-dhr2b" event={"ID":"407d1688-f160-4ee8-934d-9c76f353e4b9","Type":"ContainerStarted","Data":"3b3f4496655d256aa69efd6c02516b7d3cd50af8184669c8d8646af803224ac0"} Oct 02 13:08:52 crc kubenswrapper[4710]: E1002 13:08:52.643965 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-openstackclient:watcher_latest" Oct 02 13:08:52 crc kubenswrapper[4710]: E1002 13:08:52.644015 4710 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.147:5001/podified-master-centos10/openstack-openstackclient:watcher_latest" Oct 02 13:08:52 crc kubenswrapper[4710]: E1002 13:08:52.644127 4710 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:38.102.83.147:5001/podified-master-centos10/openstack-openstackclient:watcher_latest,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n678h64dh5c4h9bh66dh5f8h58ch5c7h65ch68fh67ch574h566h5b9h668h67bh589h549h548h5ch555h558h55h66bhfch59ch569h64dhbch675h5c9hb9q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrc4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(71a2a3d5-5a6a-4f31-8df3-f087529d550d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 13:08:52 crc kubenswrapper[4710]: E1002 13:08:52.645459 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="71a2a3d5-5a6a-4f31-8df3-f087529d550d" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.653037 4710 generic.go:334] "Generic (PLEG): container finished" podID="27670d96-0e34-41ae-a54c-2bb7f19b1c6e" containerID="35739b19eacea9fcf2439673fc3c3a190f8a8a108b09dc5fe1742feecb333446" exitCode=2 Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.653128 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"27670d96-0e34-41ae-a54c-2bb7f19b1c6e","Type":"ContainerDied","Data":"35739b19eacea9fcf2439673fc3c3a190f8a8a108b09dc5fe1742feecb333446"} Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.670470 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f85867556-bdslh" event={"ID":"00020afa-1e01-4a2c-924e-2aa55bac0e85","Type":"ContainerDied","Data":"96556850eff82ad42e377905b1f71da4a72c48cac826b297323ecf1b5f654856"} Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.670535 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96556850eff82ad42e377905b1f71da4a72c48cac826b297323ecf1b5f654856" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.741129 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.746984 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.785175 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.921823 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data-custom\") pod \"770412a4-b41f-4d11-986c-f585fae01d7c\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.922126 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-secret-key\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.922512 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-tls-certs\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.922646 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-sg-core-conf-yaml\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.922824 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00020afa-1e01-4a2c-924e-2aa55bac0e85-logs\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.922929 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-scripts\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.923217 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/770412a4-b41f-4d11-986c-f585fae01d7c-etc-machine-id\") pod \"770412a4-b41f-4d11-986c-f585fae01d7c\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.923392 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data\") pod \"770412a4-b41f-4d11-986c-f585fae01d7c\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.923732 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-config-data\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.925254 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-combined-ca-bundle\") pod \"770412a4-b41f-4d11-986c-f585fae01d7c\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.928420 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-combined-ca-bundle\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.928880 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-scripts\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929163 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b96vg\" (UniqueName: \"kubernetes.io/projected/7dc3ecad-515a-40b5-ac68-b19c0d276caa-kube-api-access-b96vg\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929210 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-scripts\") pod \"770412a4-b41f-4d11-986c-f585fae01d7c\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929327 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc9mk\" (UniqueName: \"kubernetes.io/projected/770412a4-b41f-4d11-986c-f585fae01d7c-kube-api-access-xc9mk\") pod \"770412a4-b41f-4d11-986c-f585fae01d7c\" (UID: \"770412a4-b41f-4d11-986c-f585fae01d7c\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929374 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-log-httpd\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929437 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9krh2\" (UniqueName: \"kubernetes.io/projected/00020afa-1e01-4a2c-924e-2aa55bac0e85-kube-api-access-9krh2\") pod \"00020afa-1e01-4a2c-924e-2aa55bac0e85\" (UID: \"00020afa-1e01-4a2c-924e-2aa55bac0e85\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929485 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-combined-ca-bundle\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929527 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-run-httpd\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.929590 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-config-data\") pod \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\" (UID: \"7dc3ecad-515a-40b5-ac68-b19c0d276caa\") " Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.935410 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/770412a4-b41f-4d11-986c-f585fae01d7c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "770412a4-b41f-4d11-986c-f585fae01d7c" (UID: "770412a4-b41f-4d11-986c-f585fae01d7c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.937131 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.941762 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00020afa-1e01-4a2c-924e-2aa55bac0e85-logs" (OuterVolumeSpecName: "logs") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.942522 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.946316 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "770412a4-b41f-4d11-986c-f585fae01d7c" (UID: "770412a4-b41f-4d11-986c-f585fae01d7c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.958028 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-scripts" (OuterVolumeSpecName: "scripts") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.958912 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00020afa-1e01-4a2c-924e-2aa55bac0e85-kube-api-access-9krh2" (OuterVolumeSpecName: "kube-api-access-9krh2") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "kube-api-access-9krh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.960251 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc3ecad-515a-40b5-ac68-b19c0d276caa-kube-api-access-b96vg" (OuterVolumeSpecName: "kube-api-access-b96vg") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "kube-api-access-b96vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.965811 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.966331 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770412a4-b41f-4d11-986c-f585fae01d7c-kube-api-access-xc9mk" (OuterVolumeSpecName: "kube-api-access-xc9mk") pod "770412a4-b41f-4d11-986c-f585fae01d7c" (UID: "770412a4-b41f-4d11-986c-f585fae01d7c"). InnerVolumeSpecName "kube-api-access-xc9mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:52 crc kubenswrapper[4710]: I1002 13:08:52.979889 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-scripts" (OuterVolumeSpecName: "scripts") pod "770412a4-b41f-4d11-986c-f585fae01d7c" (UID: "770412a4-b41f-4d11-986c-f585fae01d7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.000925 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.013960 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-config-data" (OuterVolumeSpecName: "config-data") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.030325 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "770412a4-b41f-4d11-986c-f585fae01d7c" (UID: "770412a4-b41f-4d11-986c-f585fae01d7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.030787 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034131 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc9mk\" (UniqueName: \"kubernetes.io/projected/770412a4-b41f-4d11-986c-f585fae01d7c-kube-api-access-xc9mk\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034167 4710 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034176 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9krh2\" (UniqueName: \"kubernetes.io/projected/00020afa-1e01-4a2c-924e-2aa55bac0e85-kube-api-access-9krh2\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034185 4710 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dc3ecad-515a-40b5-ac68-b19c0d276caa-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034193 4710 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034203 4710 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034212 4710 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034222 4710 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034231 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00020afa-1e01-4a2c-924e-2aa55bac0e85-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034242 4710 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/770412a4-b41f-4d11-986c-f585fae01d7c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034252 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034264 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034273 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034283 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b96vg\" (UniqueName: \"kubernetes.io/projected/7dc3ecad-515a-40b5-ac68-b19c0d276caa-kube-api-access-b96vg\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.034292 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.043801 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.044488 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-scripts" (OuterVolumeSpecName: "scripts") pod "00020afa-1e01-4a2c-924e-2aa55bac0e85" (UID: "00020afa-1e01-4a2c-924e-2aa55bac0e85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.107811 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.116961 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data" (OuterVolumeSpecName: "config-data") pod "770412a4-b41f-4d11-986c-f585fae01d7c" (UID: "770412a4-b41f-4d11-986c-f585fae01d7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.136082 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.136118 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00020afa-1e01-4a2c-924e-2aa55bac0e85-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.136131 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770412a4-b41f-4d11-986c-f585fae01d7c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.136141 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00020afa-1e01-4a2c-924e-2aa55bac0e85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.153896 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-config-data" (OuterVolumeSpecName: "config-data") pod "7dc3ecad-515a-40b5-ac68-b19c0d276caa" (UID: "7dc3ecad-515a-40b5-ac68-b19c0d276caa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.238049 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc3ecad-515a-40b5-ac68-b19c0d276caa-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.682282 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"770412a4-b41f-4d11-986c-f585fae01d7c","Type":"ContainerDied","Data":"ff9816cfc65a2a4aa625b3f6d446ab7e755382bd0b3fd41372444230f00465f1"} Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.682337 4710 scope.go:117] "RemoveContainer" containerID="3ef4f270e7b999b7af90be362bdfaff814a78345b72e8869d4bdf342f9ce8966" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.682492 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.687960 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f85867556-bdslh" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.691233 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.691303 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dc3ecad-515a-40b5-ac68-b19c0d276caa","Type":"ContainerDied","Data":"244a92e59f11e6a642c2aa8e7b8181582366221f843e6a39f5052812d24568ca"} Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.697464 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.147:5001/podified-master-centos10/openstack-openstackclient:watcher_latest\\\"\"" pod="openstack/openstackclient" podUID="71a2a3d5-5a6a-4f31-8df3-f087529d550d" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.753120 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.768982 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.780902 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.790498 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.798796 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f85867556-bdslh"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.807258 4710 scope.go:117] "RemoveContainer" containerID="71633b08c3b5aae4bac4e04d2ea2ad8365b518d22b20aa87db560d45369cfb09" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.809685 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.810203 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="probe" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.810283 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="probe" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.810366 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="cinder-scheduler" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.810446 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="cinder-scheduler" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.810519 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-central-agent" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.810578 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-central-agent" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.810651 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon-log" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.810714 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon-log" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.810800 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="sg-core" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.810884 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="sg-core" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.810969 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-notification-agent" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811035 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-notification-agent" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.811110 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811176 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" Oct 02 13:08:53 crc kubenswrapper[4710]: E1002 13:08:53.811250 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="proxy-httpd" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811309 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="proxy-httpd" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811582 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811667 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" containerName="horizon-log" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811738 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-notification-agent" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811893 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="cinder-scheduler" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.811967 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="proxy-httpd" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.812042 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="sg-core" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.812203 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" containerName="probe" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.812285 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" containerName="ceilometer-central-agent" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.813518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.816712 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.816888 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.820249 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.823634 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.826266 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f85867556-bdslh"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.827291 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.837247 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.840565 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.844924 4710 scope.go:117] "RemoveContainer" containerID="73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847707 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847773 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847813 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-scripts\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847843 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7zmt\" (UniqueName: \"kubernetes.io/projected/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-kube-api-access-r7zmt\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847869 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847928 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847960 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-run-httpd\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.847991 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf46l\" (UniqueName: \"kubernetes.io/projected/7d5a391c-014b-4513-95f8-04975497db39-kube-api-access-gf46l\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.848047 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-scripts\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.848072 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.848092 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-log-httpd\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.848128 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-config-data\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.848143 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-config-data\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.870884 4710 scope.go:117] "RemoveContainer" containerID="bf810ce4d53e2f76e3258d11df09631564767962abf0f8557e31cc29e316cf32" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950054 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-scripts\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950332 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950353 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-log-httpd\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950382 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-config-data\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950400 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-config-data\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950421 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950438 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950461 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-scripts\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950480 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7zmt\" (UniqueName: \"kubernetes.io/projected/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-kube-api-access-r7zmt\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950496 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950535 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950569 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-run-httpd\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950612 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf46l\" (UniqueName: \"kubernetes.io/projected/7d5a391c-014b-4513-95f8-04975497db39-kube-api-access-gf46l\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:53 crc kubenswrapper[4710]: I1002 13:08:53.950868 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.270959 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-log-httpd\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.270979 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-run-httpd\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.275035 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-scripts\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.275460 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-config-data\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.276283 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.276354 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.276547 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-scripts\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.276828 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-config-data\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.276863 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.277678 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.277891 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf46l\" (UniqueName: \"kubernetes.io/projected/7d5a391c-014b-4513-95f8-04975497db39-kube-api-access-gf46l\") pod \"ceilometer-0\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.288702 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7zmt\" (UniqueName: \"kubernetes.io/projected/2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9-kube-api-access-r7zmt\") pod \"cinder-scheduler-0\" (UID: \"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9\") " pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.302216 4710 scope.go:117] "RemoveContainer" containerID="a7572dc21b0e3541f4fd4ee817688f46e16ed0b9c70d512f30df4b8982ca8772" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.457963 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.458782 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.700939 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-76f964d7cf-dhr2b" event={"ID":"407d1688-f160-4ee8-934d-9c76f353e4b9","Type":"ContainerStarted","Data":"53b7419bac7c140379dcdae308fa6566d6b98bb89b429a488facc50bc4a33dfd"} Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.704482 4710 scope.go:117] "RemoveContainer" containerID="b4d999f5769fbff464e6d35ac58c689edf5e54b43e48465dea610111f633ef4e" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.923691 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00020afa-1e01-4a2c-924e-2aa55bac0e85" path="/var/lib/kubelet/pods/00020afa-1e01-4a2c-924e-2aa55bac0e85/volumes" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.924600 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770412a4-b41f-4d11-986c-f585fae01d7c" path="/var/lib/kubelet/pods/770412a4-b41f-4d11-986c-f585fae01d7c/volumes" Oct 02 13:08:54 crc kubenswrapper[4710]: I1002 13:08:54.926201 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc3ecad-515a-40b5-ac68-b19c0d276caa" path="/var/lib/kubelet/pods/7dc3ecad-515a-40b5-ac68-b19c0d276caa/volumes" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.174421 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.223079 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.335695 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.370612 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54fm2\" (UniqueName: \"kubernetes.io/projected/27670d96-0e34-41ae-a54c-2bb7f19b1c6e-kube-api-access-54fm2\") pod \"27670d96-0e34-41ae-a54c-2bb7f19b1c6e\" (UID: \"27670d96-0e34-41ae-a54c-2bb7f19b1c6e\") " Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.375916 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27670d96-0e34-41ae-a54c-2bb7f19b1c6e-kube-api-access-54fm2" (OuterVolumeSpecName: "kube-api-access-54fm2") pod "27670d96-0e34-41ae-a54c-2bb7f19b1c6e" (UID: "27670d96-0e34-41ae-a54c-2bb7f19b1c6e"). InnerVolumeSpecName "kube-api-access-54fm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.473258 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54fm2\" (UniqueName: \"kubernetes.io/projected/27670d96-0e34-41ae-a54c-2bb7f19b1c6e-kube-api-access-54fm2\") on node \"crc\" DevicePath \"\"" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.710283 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerStarted","Data":"e6619bc492b888acd40cced7e146de43b5e6f0e70389420974626c40dbd3bd2e"} Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.715908 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-76f964d7cf-dhr2b" event={"ID":"407d1688-f160-4ee8-934d-9c76f353e4b9","Type":"ContainerStarted","Data":"67e791e50f9568db0e86cce9a618e247f634f29a79dc6f7781c886e3184175f8"} Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.716037 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.721004 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.721138 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"27670d96-0e34-41ae-a54c-2bb7f19b1c6e","Type":"ContainerDied","Data":"6a37868078759f9356052880e4f52e77085a7da9dd5d161981019de284561ea1"} Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.721197 4710 scope.go:117] "RemoveContainer" containerID="35739b19eacea9fcf2439673fc3c3a190f8a8a108b09dc5fe1742feecb333446" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.735876 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9","Type":"ContainerStarted","Data":"587d4b19e9ebb243407ac3423589516e1aeb0dfcd083a35ee1a24873e4ed7ef7"} Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.742206 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-76f964d7cf-dhr2b" podStartSLOduration=11.74218082 podStartE2EDuration="11.74218082s" podCreationTimestamp="2025-10-02 13:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:08:55.73550139 +0000 UTC m=+1179.841912273" watchObservedRunningTime="2025-10-02 13:08:55.74218082 +0000 UTC m=+1179.848591703" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.790287 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.803001 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.811463 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:08:55 crc kubenswrapper[4710]: E1002 13:08:55.812039 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27670d96-0e34-41ae-a54c-2bb7f19b1c6e" containerName="kube-state-metrics" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.812064 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="27670d96-0e34-41ae-a54c-2bb7f19b1c6e" containerName="kube-state-metrics" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.812235 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="27670d96-0e34-41ae-a54c-2bb7f19b1c6e" containerName="kube-state-metrics" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.813044 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.819670 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.833092 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.833614 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.995322 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.995451 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.995499 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvh2t\" (UniqueName: \"kubernetes.io/projected/40732cb5-9326-4916-91dc-21095ecdbca7-kube-api-access-nvh2t\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:55 crc kubenswrapper[4710]: I1002 13:08:55.995531 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.096945 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvh2t\" (UniqueName: \"kubernetes.io/projected/40732cb5-9326-4916-91dc-21095ecdbca7-kube-api-access-nvh2t\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.097283 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.097366 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.097450 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.101803 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.102314 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.111595 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40732cb5-9326-4916-91dc-21095ecdbca7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.118890 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvh2t\" (UniqueName: \"kubernetes.io/projected/40732cb5-9326-4916-91dc-21095ecdbca7-kube-api-access-nvh2t\") pod \"kube-state-metrics-0\" (UID: \"40732cb5-9326-4916-91dc-21095ecdbca7\") " pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.193264 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.693039 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.754128 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"40732cb5-9326-4916-91dc-21095ecdbca7","Type":"ContainerStarted","Data":"3390509bdb6ef2a95ca2bae8dc3f59b8ac8034564adb353f2ac8816416158c45"} Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.754255 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.769264 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:08:56 crc kubenswrapper[4710]: I1002 13:08:56.923984 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27670d96-0e34-41ae-a54c-2bb7f19b1c6e" path="/var/lib/kubelet/pods/27670d96-0e34-41ae-a54c-2bb7f19b1c6e/volumes" Oct 02 13:08:57 crc kubenswrapper[4710]: I1002 13:08:57.764711 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9","Type":"ContainerStarted","Data":"c5bb17291059e426005e1abf6f9e02b5890b23b9d9d5c078156305db859be828"} Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:08:59.445396 4710 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/4dd7f55580a62e205e3612945b510f6b28ab94751351b5f7d48c7851ac1a97ca/diff" to get inode usage: stat /var/lib/containers/storage/overlay/4dd7f55580a62e205e3612945b510f6b28ab94751351b5f7d48c7851ac1a97ca/diff: no such file or directory, extraDiskErr: Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:08:59.794981 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9","Type":"ContainerStarted","Data":"9530d419a6151b3552b0057b792dc027a742585d2c95ffff85f863a33a0f020e"} Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:08:59.819266 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-76f964d7cf-dhr2b" podUID="407d1688-f160-4ee8-934d-9c76f353e4b9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:08:59.819329 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-76f964d7cf-dhr2b" podUID="407d1688-f160-4ee8-934d-9c76f353e4b9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:09:00.144647 4710 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/3fc0f8c2c16874133751c511df453e2bd3da975c9ff8b3613580ccd2bc36152d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/3fc0f8c2c16874133751c511df453e2bd3da975c9ff8b3613580ccd2bc36152d/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-f5fcfd868-8dmb9_e2022f45-9ec9-483a-bce6-85945738da7f/neutron-api/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-f5fcfd868-8dmb9_e2022f45-9ec9-483a-bce6-85945738da7f/neutron-api/0.log: no such file or directory Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:00.839976 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.83995144 podStartE2EDuration="7.83995144s" podCreationTimestamp="2025-10-02 13:08:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:09:00.826917449 +0000 UTC m=+1184.933328332" watchObservedRunningTime="2025-10-02 13:09:00.83995144 +0000 UTC m=+1184.946362343" Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:09:00.894419 4710 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/0e01048356f1e11ada7924f1795b9329ff8e54e13f61122ac6cfa02d3fc23b3f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/0e01048356f1e11ada7924f1795b9329ff8e54e13f61122ac6cfa02d3fc23b3f/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-f5fcfd868-8dmb9_e2022f45-9ec9-483a-bce6-85945738da7f/neutron-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-f5fcfd868-8dmb9_e2022f45-9ec9-483a-bce6-85945738da7f/neutron-httpd/0.log: no such file or directory Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:09:01.136514 4710 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/b18aa7b214f88ce845aa96d0a47826829a5b03fbe4cb382c4c65ccddbf09b5a7/diff" to get inode usage: stat /var/lib/containers/storage/overlay/b18aa7b214f88ce845aa96d0a47826829a5b03fbe4cb382c4c65ccddbf09b5a7/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-59b4d46b6c-l4bbj_48923995-d4ba-4562-8db2-26d5d34a0be5/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-59b4d46b6c-l4bbj_48923995-d4ba-4562-8db2-26d5d34a0be5/dnsmasq-dns/0.log: no such file or directory Oct 02 13:09:05 crc kubenswrapper[4710]: W1002 13:09:03.224886 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dc3ecad_515a_40b5_ac68_b19c0d276caa.slice/crio-73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7.scope WatchSource:0}: Error finding container 73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7: Status 404 returned error can't find the container with id 73062b6020e6351b27b9070f381357f8681ef15ea0ea8bb5e801bff4323aede7 Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:03.850955 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerStarted","Data":"2097a0dd67e616851faffad75e4844fbca0daaf08339bf84a15a135e87f80a0c"} Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:03.905233 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:09:03.905466 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:04.458532 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:04.615072 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:04.818949 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:04.826332 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-76f964d7cf-dhr2b" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:04.881031 4710 generic.go:334] "Generic (PLEG): container finished" podID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerID="1c1314922533bec141ed8d13d3af2eedc05cae2eadb250a7f7776c45d549b771" exitCode=137 Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:04.881385 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"02ac8fc8-b3db-4908-9322-ff4318f21a82","Type":"ContainerDied","Data":"1c1314922533bec141ed8d13d3af2eedc05cae2eadb250a7f7776c45d549b771"} Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.081072 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181703 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181790 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57mvw\" (UniqueName: \"kubernetes.io/projected/02ac8fc8-b3db-4908-9322-ff4318f21a82-kube-api-access-57mvw\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181813 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-combined-ca-bundle\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181832 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ac8fc8-b3db-4908-9322-ff4318f21a82-etc-machine-id\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181890 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-scripts\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181912 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data-custom\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.181976 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02ac8fc8-b3db-4908-9322-ff4318f21a82-logs\") pod \"02ac8fc8-b3db-4908-9322-ff4318f21a82\" (UID: \"02ac8fc8-b3db-4908-9322-ff4318f21a82\") " Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.182575 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ac8fc8-b3db-4908-9322-ff4318f21a82-logs" (OuterVolumeSpecName: "logs") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.182626 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02ac8fc8-b3db-4908-9322-ff4318f21a82-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.188509 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-scripts" (OuterVolumeSpecName: "scripts") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.189111 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.189282 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ac8fc8-b3db-4908-9322-ff4318f21a82-kube-api-access-57mvw" (OuterVolumeSpecName: "kube-api-access-57mvw") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "kube-api-access-57mvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.216286 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.240841 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data" (OuterVolumeSpecName: "config-data") pod "02ac8fc8-b3db-4908-9322-ff4318f21a82" (UID: "02ac8fc8-b3db-4908-9322-ff4318f21a82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284656 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02ac8fc8-b3db-4908-9322-ff4318f21a82-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284686 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284698 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57mvw\" (UniqueName: \"kubernetes.io/projected/02ac8fc8-b3db-4908-9322-ff4318f21a82-kube-api-access-57mvw\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284709 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284768 4710 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ac8fc8-b3db-4908-9322-ff4318f21a82-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284776 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.284784 4710 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02ac8fc8-b3db-4908-9322-ff4318f21a82-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.892433 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"02ac8fc8-b3db-4908-9322-ff4318f21a82","Type":"ContainerDied","Data":"9e68648c3fae9a0324664bc73173a1332f93071cdf6d97e5dee19b42f0f73657"} Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.892497 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.892508 4710 scope.go:117] "RemoveContainer" containerID="1c1314922533bec141ed8d13d3af2eedc05cae2eadb250a7f7776c45d549b771" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.940020 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.952988 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.959102 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:09:05.959472 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api-log" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.959491 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api-log" Oct 02 13:09:05 crc kubenswrapper[4710]: E1002 13:09:05.959522 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.959530 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.959704 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.959731 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" containerName="cinder-api-log" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.962252 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.965608 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.965836 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.965968 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 13:09:05 crc kubenswrapper[4710]: I1002 13:09:05.990038 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101577 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101644 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101672 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101760 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101829 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101863 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-config-data\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101940 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4l2l\" (UniqueName: \"kubernetes.io/projected/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-kube-api-access-g4l2l\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.101997 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-logs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.102019 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-scripts\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.203888 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204023 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204062 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204102 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-config-data\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204184 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4l2l\" (UniqueName: \"kubernetes.io/projected/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-kube-api-access-g4l2l\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204254 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-logs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204279 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-scripts\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204322 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.204362 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.207283 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.207940 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-logs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.209856 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.211159 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-config-data\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.215036 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.217390 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.217618 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.217706 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-scripts\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.222195 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4l2l\" (UniqueName: \"kubernetes.io/projected/4f909e63-ebbe-4888-ace1-0f9595d8b5cc-kube-api-access-g4l2l\") pod \"cinder-api-0\" (UID: \"4f909e63-ebbe-4888-ace1-0f9595d8b5cc\") " pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.296164 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.364996 4710 scope.go:117] "RemoveContainer" containerID="b3d79be00a538bc5e49dcc2998708519fe59a6073e4e64331ed8a88afaa78853" Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.849871 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 13:09:06 crc kubenswrapper[4710]: W1002 13:09:06.852166 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f909e63_ebbe_4888_ace1_0f9595d8b5cc.slice/crio-12af6a8be55bd9b23c045b27d0e63d6c4b08809642e3f26fa01143b8ca4c69e3 WatchSource:0}: Error finding container 12af6a8be55bd9b23c045b27d0e63d6c4b08809642e3f26fa01143b8ca4c69e3: Status 404 returned error can't find the container with id 12af6a8be55bd9b23c045b27d0e63d6c4b08809642e3f26fa01143b8ca4c69e3 Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.903327 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4f909e63-ebbe-4888-ace1-0f9595d8b5cc","Type":"ContainerStarted","Data":"12af6a8be55bd9b23c045b27d0e63d6c4b08809642e3f26fa01143b8ca4c69e3"} Oct 02 13:09:06 crc kubenswrapper[4710]: I1002 13:09:06.925998 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ac8fc8-b3db-4908-9322-ff4318f21a82" path="/var/lib/kubelet/pods/02ac8fc8-b3db-4908-9322-ff4318f21a82/volumes" Oct 02 13:09:07 crc kubenswrapper[4710]: I1002 13:09:07.918425 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"40732cb5-9326-4916-91dc-21095ecdbca7","Type":"ContainerStarted","Data":"93277d49661102c539c943b44ae91f6ec1ceb736bb01510e74e95f1b044a6b47"} Oct 02 13:09:09 crc kubenswrapper[4710]: I1002 13:09:09.943906 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4f909e63-ebbe-4888-ace1-0f9595d8b5cc","Type":"ContainerStarted","Data":"c7ee182062879fe1ce3c19da8216e034f230a29ce5a8cb0ae47c567deb7bf833"} Oct 02 13:09:09 crc kubenswrapper[4710]: I1002 13:09:09.944528 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 13:09:09 crc kubenswrapper[4710]: I1002 13:09:09.970714 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=5.3006128310000005 podStartE2EDuration="14.970687557s" podCreationTimestamp="2025-10-02 13:08:55 +0000 UTC" firstStartedPulling="2025-10-02 13:08:56.695398587 +0000 UTC m=+1180.801809470" lastFinishedPulling="2025-10-02 13:09:06.365473313 +0000 UTC m=+1190.471884196" observedRunningTime="2025-10-02 13:09:09.966558302 +0000 UTC m=+1194.072969205" watchObservedRunningTime="2025-10-02 13:09:09.970687557 +0000 UTC m=+1194.077098440" Oct 02 13:09:12 crc kubenswrapper[4710]: I1002 13:09:12.092701 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:12 crc kubenswrapper[4710]: I1002 13:09:12.094300 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:12 crc kubenswrapper[4710]: I1002 13:09:12.095356 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:09:12 crc kubenswrapper[4710]: E1002 13:09:12.095729 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:09:12 crc kubenswrapper[4710]: I1002 13:09:12.971279 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:09:12 crc kubenswrapper[4710]: E1002 13:09:12.971802 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(8ecc2560-d8d5-405b-ae65-0112862e2171)\"" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" Oct 02 13:09:16 crc kubenswrapper[4710]: I1002 13:09:16.231560 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 13:09:17 crc kubenswrapper[4710]: I1002 13:09:17.491956 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:09:18 crc kubenswrapper[4710]: I1002 13:09:18.032542 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"71a2a3d5-5a6a-4f31-8df3-f087529d550d","Type":"ContainerStarted","Data":"e67f23222d1265d5f8818bece14405fd83d895ea407f589a716534a972d55324"} Oct 02 13:09:18 crc kubenswrapper[4710]: I1002 13:09:18.036694 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4f909e63-ebbe-4888-ace1-0f9595d8b5cc","Type":"ContainerStarted","Data":"d883f8be6419f1e972127a6a4917865e4afde763dda2761099f30760bd4a7143"} Oct 02 13:09:18 crc kubenswrapper[4710]: I1002 13:09:18.037517 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 13:09:18 crc kubenswrapper[4710]: I1002 13:09:18.052165 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.841813364 podStartE2EDuration="44.052146764s" podCreationTimestamp="2025-10-02 13:08:34 +0000 UTC" firstStartedPulling="2025-10-02 13:08:35.61728577 +0000 UTC m=+1159.723696653" lastFinishedPulling="2025-10-02 13:09:16.82761917 +0000 UTC m=+1200.934030053" observedRunningTime="2025-10-02 13:09:18.048320116 +0000 UTC m=+1202.154731019" watchObservedRunningTime="2025-10-02 13:09:18.052146764 +0000 UTC m=+1202.158557647" Oct 02 13:09:18 crc kubenswrapper[4710]: I1002 13:09:18.052219 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerStarted","Data":"3ea15786e97be6f4ebfcae48cc8287fb0c19f94c40fd601dd8c3a154bf2fdce7"} Oct 02 13:09:18 crc kubenswrapper[4710]: I1002 13:09:18.100739 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=13.100716998 podStartE2EDuration="13.100716998s" podCreationTimestamp="2025-10-02 13:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:09:18.07166358 +0000 UTC m=+1202.178074463" watchObservedRunningTime="2025-10-02 13:09:18.100716998 +0000 UTC m=+1202.207127881" Oct 02 13:09:20 crc kubenswrapper[4710]: I1002 13:09:20.081938 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerStarted","Data":"f3b88d5510e17174efdea5ce5208e43a27ea7f9a502068e922bf8456d835ee47"} Oct 02 13:09:21 crc kubenswrapper[4710]: I1002 13:09:21.534948 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:09:21 crc kubenswrapper[4710]: I1002 13:09:21.535620 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-httpd" containerID="cri-o://c793e42859c6f8f5a3a59a1f2c542ff0e34fcd835bde0fd3100307777c56574e" gracePeriod=30 Oct 02 13:09:21 crc kubenswrapper[4710]: I1002 13:09:21.535578 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-log" containerID="cri-o://bf7c7ff0c599541828e59be2370327c35086721fb238d5b356aeef192b120573" gracePeriod=30 Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.101926 4710 generic.go:334] "Generic (PLEG): container finished" podID="cd4716d4-8cdb-4912-9262-caca477c119a" containerID="bf7c7ff0c599541828e59be2370327c35086721fb238d5b356aeef192b120573" exitCode=143 Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.101971 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd4716d4-8cdb-4912-9262-caca477c119a","Type":"ContainerDied","Data":"bf7c7ff0c599541828e59be2370327c35086721fb238d5b356aeef192b120573"} Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.530286 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.530647 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.530703 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.531476 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51e94e7f2d48ee2b42323dae0e2a0810cf3081a0dd80ed3b534b22c393fd779b"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:09:22 crc kubenswrapper[4710]: I1002 13:09:22.531557 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://51e94e7f2d48ee2b42323dae0e2a0810cf3081a0dd80ed3b534b22c393fd779b" gracePeriod=600 Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.117284 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerStarted","Data":"f2fd45cf84eb43391ba9c6dd9d6f4bbb1e70933891c3934cb232645854594238"} Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.119430 4710 generic.go:334] "Generic (PLEG): container finished" podID="cd4716d4-8cdb-4912-9262-caca477c119a" containerID="c793e42859c6f8f5a3a59a1f2c542ff0e34fcd835bde0fd3100307777c56574e" exitCode=0 Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.119483 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd4716d4-8cdb-4912-9262-caca477c119a","Type":"ContainerDied","Data":"c793e42859c6f8f5a3a59a1f2c542ff0e34fcd835bde0fd3100307777c56574e"} Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.123951 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="51e94e7f2d48ee2b42323dae0e2a0810cf3081a0dd80ed3b534b22c393fd779b" exitCode=0 Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.124000 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"51e94e7f2d48ee2b42323dae0e2a0810cf3081a0dd80ed3b534b22c393fd779b"} Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.124061 4710 scope.go:117] "RemoveContainer" containerID="be59f3d65c42b4df4c251f325b3eef5d2ce156af0437d5494544eb88ffa0a1fd" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.163494 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.163714 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-log" containerID="cri-o://571b65a30e776db3a5c455b8f64efd6f218c1c142dad47158a753fae976a139c" gracePeriod=30 Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.164108 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-httpd" containerID="cri-o://c3e02d5842f0eff8776966942e67439672947e4cbb6d9da165b5d2a389112b31" gracePeriod=30 Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.708878 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.867845 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-logs\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.867982 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66244\" (UniqueName: \"kubernetes.io/projected/cd4716d4-8cdb-4912-9262-caca477c119a-kube-api-access-66244\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868042 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-config-data\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868106 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-public-tls-certs\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868146 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-combined-ca-bundle\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868192 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-scripts\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868243 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868323 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-httpd-run\") pod \"cd4716d4-8cdb-4912-9262-caca477c119a\" (UID: \"cd4716d4-8cdb-4912-9262-caca477c119a\") " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868570 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-logs" (OuterVolumeSpecName: "logs") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.868844 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.869613 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.882891 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-scripts" (OuterVolumeSpecName: "scripts") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.903021 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd4716d4-8cdb-4912-9262-caca477c119a-kube-api-access-66244" (OuterVolumeSpecName: "kube-api-access-66244") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "kube-api-access-66244". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.920024 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.946581 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.972198 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.972231 4710 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd4716d4-8cdb-4912-9262-caca477c119a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.972244 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66244\" (UniqueName: \"kubernetes.io/projected/cd4716d4-8cdb-4912-9262-caca477c119a-kube-api-access-66244\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.972259 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:23 crc kubenswrapper[4710]: I1002 13:09:23.972269 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.001835 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.017014 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-config-data" (OuterVolumeSpecName: "config-data") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.053918 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cd4716d4-8cdb-4912-9262-caca477c119a" (UID: "cd4716d4-8cdb-4912-9262-caca477c119a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.073966 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.074019 4710 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd4716d4-8cdb-4912-9262-caca477c119a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.074031 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.180840 4710 generic.go:334] "Generic (PLEG): container finished" podID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerID="571b65a30e776db3a5c455b8f64efd6f218c1c142dad47158a753fae976a139c" exitCode=143 Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.180943 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48daf522-43e6-4fa5-a526-ee522fda7ae2","Type":"ContainerDied","Data":"571b65a30e776db3a5c455b8f64efd6f218c1c142dad47158a753fae976a139c"} Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.201887 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"4b5183e50d1f99d425b43adb71d4ff8d51ea84c4eb7c8acc094921b9bbeee6bd"} Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.218230 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-central-agent" containerID="cri-o://2097a0dd67e616851faffad75e4844fbca0daaf08339bf84a15a135e87f80a0c" gracePeriod=30 Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.218628 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.223894 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="proxy-httpd" containerID="cri-o://f2fd45cf84eb43391ba9c6dd9d6f4bbb1e70933891c3934cb232645854594238" gracePeriod=30 Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.224080 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd4716d4-8cdb-4912-9262-caca477c119a","Type":"ContainerDied","Data":"a3fa776ca6ee57aab4a997b9f3dbf1c18b29c06a2545d570b1f22d3865487c61"} Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.224126 4710 scope.go:117] "RemoveContainer" containerID="c793e42859c6f8f5a3a59a1f2c542ff0e34fcd835bde0fd3100307777c56574e" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.224223 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.224303 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="sg-core" containerID="cri-o://f3b88d5510e17174efdea5ce5208e43a27ea7f9a502068e922bf8456d835ee47" gracePeriod=30 Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.224362 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-notification-agent" containerID="cri-o://3ea15786e97be6f4ebfcae48cc8287fb0c19f94c40fd601dd8c3a154bf2fdce7" gracePeriod=30 Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.305031 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.153157454 podStartE2EDuration="31.305011014s" podCreationTimestamp="2025-10-02 13:08:53 +0000 UTC" firstStartedPulling="2025-10-02 13:08:55.394452182 +0000 UTC m=+1179.500863065" lastFinishedPulling="2025-10-02 13:09:22.546305752 +0000 UTC m=+1206.652716625" observedRunningTime="2025-10-02 13:09:24.292196828 +0000 UTC m=+1208.398607711" watchObservedRunningTime="2025-10-02 13:09:24.305011014 +0000 UTC m=+1208.411421897" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.312623 4710 scope.go:117] "RemoveContainer" containerID="bf7c7ff0c599541828e59be2370327c35086721fb238d5b356aeef192b120573" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.334736 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.395198 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.468987 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:09:24 crc kubenswrapper[4710]: E1002 13:09:24.469479 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-log" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.469493 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-log" Oct 02 13:09:24 crc kubenswrapper[4710]: E1002 13:09:24.469545 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-httpd" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.469553 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-httpd" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.469778 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-httpd" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.469800 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" containerName="glance-log" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.471052 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.473494 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.473760 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.479422 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584062 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584121 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhvl6\" (UniqueName: \"kubernetes.io/projected/939d6b43-cd24-4648-8134-19f384e171bd-kube-api-access-mhvl6\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584196 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-config-data\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584358 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584414 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584490 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/939d6b43-cd24-4648-8134-19f384e171bd-logs\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584549 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-scripts\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.584680 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/939d6b43-cd24-4648-8134-19f384e171bd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.685885 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhvl6\" (UniqueName: \"kubernetes.io/projected/939d6b43-cd24-4648-8134-19f384e171bd-kube-api-access-mhvl6\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.685987 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-config-data\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686036 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686056 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686090 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/939d6b43-cd24-4648-8134-19f384e171bd-logs\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686117 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-scripts\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686150 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/939d6b43-cd24-4648-8134-19f384e171bd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686174 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686392 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686627 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/939d6b43-cd24-4648-8134-19f384e171bd-logs\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.686662 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/939d6b43-cd24-4648-8134-19f384e171bd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.691893 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.692025 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-scripts\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.692690 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-config-data\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.699606 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/939d6b43-cd24-4648-8134-19f384e171bd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.706655 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhvl6\" (UniqueName: \"kubernetes.io/projected/939d6b43-cd24-4648-8134-19f384e171bd-kube-api-access-mhvl6\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.716293 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"939d6b43-cd24-4648-8134-19f384e171bd\") " pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.786521 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 13:09:24 crc kubenswrapper[4710]: I1002 13:09:24.926214 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd4716d4-8cdb-4912-9262-caca477c119a" path="/var/lib/kubelet/pods/cd4716d4-8cdb-4912-9262-caca477c119a/volumes" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239142 4710 generic.go:334] "Generic (PLEG): container finished" podID="7d5a391c-014b-4513-95f8-04975497db39" containerID="f2fd45cf84eb43391ba9c6dd9d6f4bbb1e70933891c3934cb232645854594238" exitCode=0 Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239178 4710 generic.go:334] "Generic (PLEG): container finished" podID="7d5a391c-014b-4513-95f8-04975497db39" containerID="f3b88d5510e17174efdea5ce5208e43a27ea7f9a502068e922bf8456d835ee47" exitCode=2 Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239185 4710 generic.go:334] "Generic (PLEG): container finished" podID="7d5a391c-014b-4513-95f8-04975497db39" containerID="3ea15786e97be6f4ebfcae48cc8287fb0c19f94c40fd601dd8c3a154bf2fdce7" exitCode=0 Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239190 4710 generic.go:334] "Generic (PLEG): container finished" podID="7d5a391c-014b-4513-95f8-04975497db39" containerID="2097a0dd67e616851faffad75e4844fbca0daaf08339bf84a15a135e87f80a0c" exitCode=0 Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239230 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerDied","Data":"f2fd45cf84eb43391ba9c6dd9d6f4bbb1e70933891c3934cb232645854594238"} Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239259 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerDied","Data":"f3b88d5510e17174efdea5ce5208e43a27ea7f9a502068e922bf8456d835ee47"} Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239268 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerDied","Data":"3ea15786e97be6f4ebfcae48cc8287fb0c19f94c40fd601dd8c3a154bf2fdce7"} Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.239278 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerDied","Data":"2097a0dd67e616851faffad75e4844fbca0daaf08339bf84a15a135e87f80a0c"} Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.246285 4710 generic.go:334] "Generic (PLEG): container finished" podID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerID="c3e02d5842f0eff8776966942e67439672947e4cbb6d9da165b5d2a389112b31" exitCode=0 Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.247358 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48daf522-43e6-4fa5-a526-ee522fda7ae2","Type":"ContainerDied","Data":"c3e02d5842f0eff8776966942e67439672947e4cbb6d9da165b5d2a389112b31"} Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.432035 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 13:09:25 crc kubenswrapper[4710]: W1002 13:09:25.469278 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod939d6b43_cd24_4648_8134_19f384e171bd.slice/crio-a63b1b7091b15d41c934cf3282ad879dff4bbd50d1738d10e48a38b70a5a79dd WatchSource:0}: Error finding container a63b1b7091b15d41c934cf3282ad879dff4bbd50d1738d10e48a38b70a5a79dd: Status 404 returned error can't find the container with id a63b1b7091b15d41c934cf3282ad879dff4bbd50d1738d10e48a38b70a5a79dd Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.623638 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.716602 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbk4v\" (UniqueName: \"kubernetes.io/projected/48daf522-43e6-4fa5-a526-ee522fda7ae2-kube-api-access-mbk4v\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.716760 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-scripts\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.716809 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-internal-tls-certs\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.716833 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-config-data\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.716910 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-httpd-run\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.717004 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.717071 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-logs\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.717116 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-combined-ca-bundle\") pod \"48daf522-43e6-4fa5-a526-ee522fda7ae2\" (UID: \"48daf522-43e6-4fa5-a526-ee522fda7ae2\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.717802 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-logs" (OuterVolumeSpecName: "logs") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.718054 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.718184 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.728547 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-scripts" (OuterVolumeSpecName: "scripts") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.728628 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48daf522-43e6-4fa5-a526-ee522fda7ae2-kube-api-access-mbk4v" (OuterVolumeSpecName: "kube-api-access-mbk4v") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "kube-api-access-mbk4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.741486 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.793942 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.800302 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.815562 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-config-data" (OuterVolumeSpecName: "config-data") pod "48daf522-43e6-4fa5-a526-ee522fda7ae2" (UID: "48daf522-43e6-4fa5-a526-ee522fda7ae2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821150 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821183 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbk4v\" (UniqueName: \"kubernetes.io/projected/48daf522-43e6-4fa5-a526-ee522fda7ae2-kube-api-access-mbk4v\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821195 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821206 4710 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821213 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48daf522-43e6-4fa5-a526-ee522fda7ae2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821221 4710 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48daf522-43e6-4fa5-a526-ee522fda7ae2-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.821251 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.822040 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.888283 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922004 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-run-httpd\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922062 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-log-httpd\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922094 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf46l\" (UniqueName: \"kubernetes.io/projected/7d5a391c-014b-4513-95f8-04975497db39-kube-api-access-gf46l\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922233 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-scripts\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922469 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-combined-ca-bundle\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922595 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922869 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-sg-core-conf-yaml\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.922931 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-config-data\") pod \"7d5a391c-014b-4513-95f8-04975497db39\" (UID: \"7d5a391c-014b-4513-95f8-04975497db39\") " Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.923384 4710 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.923387 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.923400 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.947655 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5a391c-014b-4513-95f8-04975497db39-kube-api-access-gf46l" (OuterVolumeSpecName: "kube-api-access-gf46l") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "kube-api-access-gf46l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:25 crc kubenswrapper[4710]: I1002 13:09:25.970017 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-scripts" (OuterVolumeSpecName: "scripts") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.010372 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.026800 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.026839 4710 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.026851 4710 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d5a391c-014b-4513-95f8-04975497db39-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.026861 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf46l\" (UniqueName: \"kubernetes.io/projected/7d5a391c-014b-4513-95f8-04975497db39-kube-api-access-gf46l\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.079942 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.128950 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.134406 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-config-data" (OuterVolumeSpecName: "config-data") pod "7d5a391c-014b-4513-95f8-04975497db39" (UID: "7d5a391c-014b-4513-95f8-04975497db39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.233114 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d5a391c-014b-4513-95f8-04975497db39-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.268521 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48daf522-43e6-4fa5-a526-ee522fda7ae2","Type":"ContainerDied","Data":"d76fb019c7a67cdccccdc7bd92f172efd406ed630190616d3565233ead6366cf"} Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.268578 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.268592 4710 scope.go:117] "RemoveContainer" containerID="c3e02d5842f0eff8776966942e67439672947e4cbb6d9da165b5d2a389112b31" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.271184 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"939d6b43-cd24-4648-8134-19f384e171bd","Type":"ContainerStarted","Data":"a63b1b7091b15d41c934cf3282ad879dff4bbd50d1738d10e48a38b70a5a79dd"} Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.276126 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d5a391c-014b-4513-95f8-04975497db39","Type":"ContainerDied","Data":"e6619bc492b888acd40cced7e146de43b5e6f0e70389420974626c40dbd3bd2e"} Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.276246 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.305323 4710 scope.go:117] "RemoveContainer" containerID="571b65a30e776db3a5c455b8f64efd6f218c1c142dad47158a753fae976a139c" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.332152 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.346078 4710 scope.go:117] "RemoveContainer" containerID="f2fd45cf84eb43391ba9c6dd9d6f4bbb1e70933891c3934cb232645854594238" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.359353 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.368826 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.389991 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: E1002 13:09:26.390623 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-central-agent" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.390644 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-central-agent" Oct 02 13:09:26 crc kubenswrapper[4710]: E1002 13:09:26.390673 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-notification-agent" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.390684 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-notification-agent" Oct 02 13:09:26 crc kubenswrapper[4710]: E1002 13:09:26.390699 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-httpd" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.390707 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-httpd" Oct 02 13:09:26 crc kubenswrapper[4710]: E1002 13:09:26.390721 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-log" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.390728 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-log" Oct 02 13:09:26 crc kubenswrapper[4710]: E1002 13:09:26.390763 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="proxy-httpd" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.390801 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="proxy-httpd" Oct 02 13:09:26 crc kubenswrapper[4710]: E1002 13:09:26.390815 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="sg-core" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.390825 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="sg-core" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.391036 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="sg-core" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.391049 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="proxy-httpd" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.391064 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-notification-agent" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.391072 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-httpd" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.391091 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5a391c-014b-4513-95f8-04975497db39" containerName="ceilometer-central-agent" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.391101 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" containerName="glance-log" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.392832 4710 scope.go:117] "RemoveContainer" containerID="f3b88d5510e17174efdea5ce5208e43a27ea7f9a502068e922bf8456d835ee47" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.415985 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.416173 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.422172 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.422396 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.422558 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.422661 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.447936 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.450026 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.453940 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.454270 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.462289 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.481131 4710 scope.go:117] "RemoveContainer" containerID="3ea15786e97be6f4ebfcae48cc8287fb0c19f94c40fd601dd8c3a154bf2fdce7" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540163 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-scripts\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540224 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-config-data\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540259 4710 scope.go:117] "RemoveContainer" containerID="2097a0dd67e616851faffad75e4844fbca0daaf08339bf84a15a135e87f80a0c" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540276 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540410 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j7gp\" (UniqueName: \"kubernetes.io/projected/47447a8c-f16f-4519-8b8a-e83c95cf32bd-kube-api-access-5j7gp\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540425 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b7cc70-507d-4e65-98c0-820912b73208-logs\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540449 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-run-httpd\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540494 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540526 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540551 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-log-httpd\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540579 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73b7cc70-507d-4e65-98c0-820912b73208-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540613 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540868 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540913 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-scripts\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.540981 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v45r5\" (UniqueName: \"kubernetes.io/projected/73b7cc70-507d-4e65-98c0-820912b73208-kube-api-access-v45r5\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.541116 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.541147 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-config-data\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.645544 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73b7cc70-507d-4e65-98c0-820912b73208-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.645991 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646034 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-scripts\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646063 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646092 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v45r5\" (UniqueName: \"kubernetes.io/projected/73b7cc70-507d-4e65-98c0-820912b73208-kube-api-access-v45r5\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646138 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-config-data\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646171 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646223 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-scripts\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646257 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-config-data\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646289 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646310 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j7gp\" (UniqueName: \"kubernetes.io/projected/47447a8c-f16f-4519-8b8a-e83c95cf32bd-kube-api-access-5j7gp\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646332 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b7cc70-507d-4e65-98c0-820912b73208-logs\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646358 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-run-httpd\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646406 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646443 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646471 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-log-httpd\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646650 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.646137 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73b7cc70-507d-4e65-98c0-820912b73208-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.647010 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-log-httpd\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.647277 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-run-httpd\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.650932 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b7cc70-507d-4e65-98c0-820912b73208-logs\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.653431 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.654397 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-config-data\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.654485 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-config-data\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.655236 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.655362 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.656227 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-scripts\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.656906 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.657403 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b7cc70-507d-4e65-98c0-820912b73208-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.659174 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-scripts\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.667950 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v45r5\" (UniqueName: \"kubernetes.io/projected/73b7cc70-507d-4e65-98c0-820912b73208-kube-api-access-v45r5\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.667959 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j7gp\" (UniqueName: \"kubernetes.io/projected/47447a8c-f16f-4519-8b8a-e83c95cf32bd-kube-api-access-5j7gp\") pod \"ceilometer-0\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.685762 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"73b7cc70-507d-4e65-98c0-820912b73208\") " pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.771222 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.782811 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.924357 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48daf522-43e6-4fa5-a526-ee522fda7ae2" path="/var/lib/kubelet/pods/48daf522-43e6-4fa5-a526-ee522fda7ae2/volumes" Oct 02 13:09:26 crc kubenswrapper[4710]: I1002 13:09:26.925413 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d5a391c-014b-4513-95f8-04975497db39" path="/var/lib/kubelet/pods/7d5a391c-014b-4513-95f8-04975497db39/volumes" Oct 02 13:09:27 crc kubenswrapper[4710]: W1002 13:09:27.301997 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47447a8c_f16f_4519_8b8a_e83c95cf32bd.slice/crio-e2a5299ff538b830ec5196d3b2f158a01c8e6e6b3f2305ad9862c439368c55e9 WatchSource:0}: Error finding container e2a5299ff538b830ec5196d3b2f158a01c8e6e6b3f2305ad9862c439368c55e9: Status 404 returned error can't find the container with id e2a5299ff538b830ec5196d3b2f158a01c8e6e6b3f2305ad9862c439368c55e9 Oct 02 13:09:27 crc kubenswrapper[4710]: I1002 13:09:27.303453 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:27 crc kubenswrapper[4710]: I1002 13:09:27.565502 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 13:09:27 crc kubenswrapper[4710]: I1002 13:09:27.905291 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:09:28 crc kubenswrapper[4710]: I1002 13:09:28.334005 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73b7cc70-507d-4e65-98c0-820912b73208","Type":"ContainerStarted","Data":"ce69e6a27573381794ef3eadcc9888ffe799920c30cff9cc228a8a52b76f3b1a"} Oct 02 13:09:28 crc kubenswrapper[4710]: I1002 13:09:28.343501 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"939d6b43-cd24-4648-8134-19f384e171bd","Type":"ContainerStarted","Data":"05fb9ed5ab8dc00758579cbbba619b5d07c96f71744d0e2481a4fe9828c8cc8a"} Oct 02 13:09:28 crc kubenswrapper[4710]: I1002 13:09:28.347154 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerStarted","Data":"e2a5299ff538b830ec5196d3b2f158a01c8e6e6b3f2305ad9862c439368c55e9"} Oct 02 13:09:28 crc kubenswrapper[4710]: I1002 13:09:28.630308 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 13:09:29 crc kubenswrapper[4710]: I1002 13:09:29.372202 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73b7cc70-507d-4e65-98c0-820912b73208","Type":"ContainerStarted","Data":"97af631305dfc927c6db8d2e3fc863da6cce2e82375bf6b80e69a9bc7bfaaaf3"} Oct 02 13:09:29 crc kubenswrapper[4710]: I1002 13:09:29.379509 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"939d6b43-cd24-4648-8134-19f384e171bd","Type":"ContainerStarted","Data":"ebeec42887cc0605fe8ce7945746b071304afde38f72b3afafb003f0618e9804"} Oct 02 13:09:29 crc kubenswrapper[4710]: I1002 13:09:29.389601 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerStarted","Data":"f6a44d20123abc92f2a214541451df0e7fa442a4297362854326e49ffb4976be"} Oct 02 13:09:29 crc kubenswrapper[4710]: I1002 13:09:29.419735 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.419716455 podStartE2EDuration="5.419716455s" podCreationTimestamp="2025-10-02 13:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:09:29.410012268 +0000 UTC m=+1213.516423161" watchObservedRunningTime="2025-10-02 13:09:29.419716455 +0000 UTC m=+1213.526127338" Oct 02 13:09:30 crc kubenswrapper[4710]: I1002 13:09:30.430247 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73b7cc70-507d-4e65-98c0-820912b73208","Type":"ContainerStarted","Data":"fd7534cdfa30b1c224e454e9378c953be65fc85710a9f511aafab9bcc49aa847"} Oct 02 13:09:30 crc kubenswrapper[4710]: I1002 13:09:30.434061 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerStarted","Data":"422a814e5f296ac2e96f77a5c34795235ac4aa7aff890438f4aa60fd38d14778"} Oct 02 13:09:31 crc kubenswrapper[4710]: I1002 13:09:31.455114 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerStarted","Data":"0dbb9a56d30b760b60878988c255801c49867d93a8d35715d17d4287e519ef3d"} Oct 02 13:09:31 crc kubenswrapper[4710]: I1002 13:09:31.693113 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.693095258 podStartE2EDuration="5.693095258s" podCreationTimestamp="2025-10-02 13:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:09:30.460281003 +0000 UTC m=+1214.566691906" watchObservedRunningTime="2025-10-02 13:09:31.693095258 +0000 UTC m=+1215.799506141" Oct 02 13:09:31 crc kubenswrapper[4710]: I1002 13:09:31.699295 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:32 crc kubenswrapper[4710]: I1002 13:09:32.092352 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:32 crc kubenswrapper[4710]: I1002 13:09:32.170713 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:32 crc kubenswrapper[4710]: I1002 13:09:32.466894 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerStarted","Data":"779f729d982683576b8816d70e45e32dfca077ab3e0dc3b6379d68329a232c96"} Oct 02 13:09:32 crc kubenswrapper[4710]: I1002 13:09:32.467774 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:32 crc kubenswrapper[4710]: I1002 13:09:32.503502 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:32 crc kubenswrapper[4710]: I1002 13:09:32.540532 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.499634 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerStarted","Data":"8de48679b0172dc6d79861bf200716fda285094394c9451324f325077887581e"} Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.500111 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.499875 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-notification-agent" containerID="cri-o://0dbb9a56d30b760b60878988c255801c49867d93a8d35715d17d4287e519ef3d" gracePeriod=30 Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.499878 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="sg-core" containerID="cri-o://779f729d982683576b8816d70e45e32dfca077ab3e0dc3b6379d68329a232c96" gracePeriod=30 Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.499895 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="proxy-httpd" containerID="cri-o://8de48679b0172dc6d79861bf200716fda285094394c9451324f325077887581e" gracePeriod=30 Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.500108 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-central-agent" containerID="cri-o://422a814e5f296ac2e96f77a5c34795235ac4aa7aff890438f4aa60fd38d14778" gracePeriod=30 Oct 02 13:09:33 crc kubenswrapper[4710]: I1002 13:09:33.528957 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.093571798 podStartE2EDuration="7.5289271s" podCreationTimestamp="2025-10-02 13:09:26 +0000 UTC" firstStartedPulling="2025-10-02 13:09:27.309347035 +0000 UTC m=+1211.415757918" lastFinishedPulling="2025-10-02 13:09:32.744702337 +0000 UTC m=+1216.851113220" observedRunningTime="2025-10-02 13:09:33.525162694 +0000 UTC m=+1217.631573567" watchObservedRunningTime="2025-10-02 13:09:33.5289271 +0000 UTC m=+1217.635337983" Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.509414 4710 generic.go:334] "Generic (PLEG): container finished" podID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerID="8de48679b0172dc6d79861bf200716fda285094394c9451324f325077887581e" exitCode=0 Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.509800 4710 generic.go:334] "Generic (PLEG): container finished" podID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerID="779f729d982683576b8816d70e45e32dfca077ab3e0dc3b6379d68329a232c96" exitCode=2 Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.509811 4710 generic.go:334] "Generic (PLEG): container finished" podID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerID="0dbb9a56d30b760b60878988c255801c49867d93a8d35715d17d4287e519ef3d" exitCode=0 Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.509484 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerDied","Data":"8de48679b0172dc6d79861bf200716fda285094394c9451324f325077887581e"} Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.509909 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerDied","Data":"779f729d982683576b8816d70e45e32dfca077ab3e0dc3b6379d68329a232c96"} Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.509921 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerDied","Data":"0dbb9a56d30b760b60878988c255801c49867d93a8d35715d17d4287e519ef3d"} Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.510015 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" containerID="cri-o://f6a44d20123abc92f2a214541451df0e7fa442a4297362854326e49ffb4976be" gracePeriod=30 Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.787581 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.788031 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.829174 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 13:09:34 crc kubenswrapper[4710]: I1002 13:09:34.833292 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.528106 4710 generic.go:334] "Generic (PLEG): container finished" podID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerID="f6a44d20123abc92f2a214541451df0e7fa442a4297362854326e49ffb4976be" exitCode=0 Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.528149 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerDied","Data":"f6a44d20123abc92f2a214541451df0e7fa442a4297362854326e49ffb4976be"} Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.528614 4710 scope.go:117] "RemoveContainer" containerID="8ffa39df85182c62c8aec7ad533edb4aa962af2fe35591975030615937d06c44" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.528957 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.528993 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.741332 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.857895 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-combined-ca-bundle\") pod \"8ecc2560-d8d5-405b-ae65-0112862e2171\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.858043 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-custom-prometheus-ca\") pod \"8ecc2560-d8d5-405b-ae65-0112862e2171\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.858064 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-config-data\") pod \"8ecc2560-d8d5-405b-ae65-0112862e2171\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.858204 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq6jv\" (UniqueName: \"kubernetes.io/projected/8ecc2560-d8d5-405b-ae65-0112862e2171-kube-api-access-xq6jv\") pod \"8ecc2560-d8d5-405b-ae65-0112862e2171\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.858225 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ecc2560-d8d5-405b-ae65-0112862e2171-logs\") pod \"8ecc2560-d8d5-405b-ae65-0112862e2171\" (UID: \"8ecc2560-d8d5-405b-ae65-0112862e2171\") " Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.858833 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ecc2560-d8d5-405b-ae65-0112862e2171-logs" (OuterVolumeSpecName: "logs") pod "8ecc2560-d8d5-405b-ae65-0112862e2171" (UID: "8ecc2560-d8d5-405b-ae65-0112862e2171"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.865496 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ecc2560-d8d5-405b-ae65-0112862e2171-kube-api-access-xq6jv" (OuterVolumeSpecName: "kube-api-access-xq6jv") pod "8ecc2560-d8d5-405b-ae65-0112862e2171" (UID: "8ecc2560-d8d5-405b-ae65-0112862e2171"). InnerVolumeSpecName "kube-api-access-xq6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.897068 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "8ecc2560-d8d5-405b-ae65-0112862e2171" (UID: "8ecc2560-d8d5-405b-ae65-0112862e2171"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.909054 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ecc2560-d8d5-405b-ae65-0112862e2171" (UID: "8ecc2560-d8d5-405b-ae65-0112862e2171"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.922072 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-config-data" (OuterVolumeSpecName: "config-data") pod "8ecc2560-d8d5-405b-ae65-0112862e2171" (UID: "8ecc2560-d8d5-405b-ae65-0112862e2171"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.959844 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq6jv\" (UniqueName: \"kubernetes.io/projected/8ecc2560-d8d5-405b-ae65-0112862e2171-kube-api-access-xq6jv\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.959874 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ecc2560-d8d5-405b-ae65-0112862e2171-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.959885 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.959895 4710 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:35 crc kubenswrapper[4710]: I1002 13:09:35.959904 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ecc2560-d8d5-405b-ae65-0112862e2171-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.539541 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.539538 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ecc2560-d8d5-405b-ae65-0112862e2171","Type":"ContainerDied","Data":"f891f4584ab0f0269dea4e6ebcf7c7e1008c21c3294c9d14603c5ca9b9e2d3f8"} Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.540020 4710 scope.go:117] "RemoveContainer" containerID="f6a44d20123abc92f2a214541451df0e7fa442a4297362854326e49ffb4976be" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.579064 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.609220 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.621641 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:09:36 crc kubenswrapper[4710]: E1002 13:09:36.622368 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.622497 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: E1002 13:09:36.622587 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.622662 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: E1002 13:09:36.622786 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.622876 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.623189 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.623298 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.623397 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.623482 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.623572 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.624491 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.627773 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.652854 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.778663 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.778723 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qp5k\" (UniqueName: \"kubernetes.io/projected/549c5440-af2a-4aa2-8d33-a28dde44a69b-kube-api-access-8qp5k\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.779104 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549c5440-af2a-4aa2-8d33-a28dde44a69b-logs\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.779230 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.779333 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-config-data\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.783878 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.783923 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.814364 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.844242 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.880969 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549c5440-af2a-4aa2-8d33-a28dde44a69b-logs\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.881056 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.881166 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-config-data\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.881301 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.881324 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qp5k\" (UniqueName: \"kubernetes.io/projected/549c5440-af2a-4aa2-8d33-a28dde44a69b-kube-api-access-8qp5k\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.881519 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549c5440-af2a-4aa2-8d33-a28dde44a69b-logs\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.886824 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-config-data\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.887264 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.888274 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549c5440-af2a-4aa2-8d33-a28dde44a69b-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.908331 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qp5k\" (UniqueName: \"kubernetes.io/projected/549c5440-af2a-4aa2-8d33-a28dde44a69b-kube-api-access-8qp5k\") pod \"watcher-decision-engine-0\" (UID: \"549c5440-af2a-4aa2-8d33-a28dde44a69b\") " pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.946408 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:36 crc kubenswrapper[4710]: I1002 13:09:36.949805 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" path="/var/lib/kubelet/pods/8ecc2560-d8d5-405b-ae65-0112862e2171/volumes" Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.464115 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.555525 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"549c5440-af2a-4aa2-8d33-a28dde44a69b","Type":"ContainerStarted","Data":"b536b672c9e26231322c40b44abe18a7e92860671fb4bab785d23ec5fa69cad4"} Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.556313 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.556338 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.690517 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.690674 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 13:09:37 crc kubenswrapper[4710]: I1002 13:09:37.793714 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 13:09:38 crc kubenswrapper[4710]: I1002 13:09:38.566075 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"549c5440-af2a-4aa2-8d33-a28dde44a69b","Type":"ContainerStarted","Data":"4eb3ff2e14da74377d483081ee232fd49dccebf86fc45c699bd5f74f9ba87ada"} Oct 02 13:09:38 crc kubenswrapper[4710]: I1002 13:09:38.590043 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.590024529 podStartE2EDuration="2.590024529s" podCreationTimestamp="2025-10-02 13:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:09:38.584616481 +0000 UTC m=+1222.691027364" watchObservedRunningTime="2025-10-02 13:09:38.590024529 +0000 UTC m=+1222.696435412" Oct 02 13:09:39 crc kubenswrapper[4710]: I1002 13:09:39.578269 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 13:09:39 crc kubenswrapper[4710]: I1002 13:09:39.578800 4710 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 13:09:39 crc kubenswrapper[4710]: I1002 13:09:39.608436 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:39 crc kubenswrapper[4710]: I1002 13:09:39.612676 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 13:09:41 crc kubenswrapper[4710]: I1002 13:09:41.597582 4710 generic.go:334] "Generic (PLEG): container finished" podID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerID="422a814e5f296ac2e96f77a5c34795235ac4aa7aff890438f4aa60fd38d14778" exitCode=0 Oct 02 13:09:41 crc kubenswrapper[4710]: I1002 13:09:41.597650 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerDied","Data":"422a814e5f296ac2e96f77a5c34795235ac4aa7aff890438f4aa60fd38d14778"} Oct 02 13:09:41 crc kubenswrapper[4710]: I1002 13:09:41.949893 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046153 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-run-httpd\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046286 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-log-httpd\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046377 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-ceilometer-tls-certs\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046412 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-sg-core-conf-yaml\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046464 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046499 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-scripts\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046531 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j7gp\" (UniqueName: \"kubernetes.io/projected/47447a8c-f16f-4519-8b8a-e83c95cf32bd-kube-api-access-5j7gp\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046565 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-combined-ca-bundle\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046650 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-config-data\") pod \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\" (UID: \"47447a8c-f16f-4519-8b8a-e83c95cf32bd\") " Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.046725 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.047291 4710 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.047313 4710 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47447a8c-f16f-4519-8b8a-e83c95cf32bd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.056591 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47447a8c-f16f-4519-8b8a-e83c95cf32bd-kube-api-access-5j7gp" (OuterVolumeSpecName: "kube-api-access-5j7gp") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "kube-api-access-5j7gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.058044 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-scripts" (OuterVolumeSpecName: "scripts") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.075416 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.109938 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.120687 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.145011 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-config-data" (OuterVolumeSpecName: "config-data") pod "47447a8c-f16f-4519-8b8a-e83c95cf32bd" (UID: "47447a8c-f16f-4519-8b8a-e83c95cf32bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.149697 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.149816 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j7gp\" (UniqueName: \"kubernetes.io/projected/47447a8c-f16f-4519-8b8a-e83c95cf32bd-kube-api-access-5j7gp\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.149912 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.149986 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.150051 4710 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.150107 4710 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47447a8c-f16f-4519-8b8a-e83c95cf32bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.614230 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47447a8c-f16f-4519-8b8a-e83c95cf32bd","Type":"ContainerDied","Data":"e2a5299ff538b830ec5196d3b2f158a01c8e6e6b3f2305ad9862c439368c55e9"} Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.614346 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.614597 4710 scope.go:117] "RemoveContainer" containerID="8de48679b0172dc6d79861bf200716fda285094394c9451324f325077887581e" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.638817 4710 scope.go:117] "RemoveContainer" containerID="779f729d982683576b8816d70e45e32dfca077ab3e0dc3b6379d68329a232c96" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.652965 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.663603 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.674540 4710 scope.go:117] "RemoveContainer" containerID="0dbb9a56d30b760b60878988c255801c49867d93a8d35715d17d4287e519ef3d" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.677633 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:42 crc kubenswrapper[4710]: E1002 13:09:42.678063 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678076 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:42 crc kubenswrapper[4710]: E1002 13:09:42.678088 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="proxy-httpd" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678095 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="proxy-httpd" Oct 02 13:09:42 crc kubenswrapper[4710]: E1002 13:09:42.678111 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-central-agent" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678119 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-central-agent" Oct 02 13:09:42 crc kubenswrapper[4710]: E1002 13:09:42.678146 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-notification-agent" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678152 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-notification-agent" Oct 02 13:09:42 crc kubenswrapper[4710]: E1002 13:09:42.678169 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="sg-core" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678175 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="sg-core" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678399 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-notification-agent" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678415 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="proxy-httpd" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678434 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="sg-core" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678444 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" containerName="ceilometer-central-agent" Oct 02 13:09:42 crc kubenswrapper[4710]: E1002 13:09:42.678630 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.678639 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecc2560-d8d5-405b-ae65-0112862e2171" containerName="watcher-decision-engine" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.683582 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.685620 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.685774 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.685939 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.693355 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.707777 4710 scope.go:117] "RemoveContainer" containerID="422a814e5f296ac2e96f77a5c34795235ac4aa7aff890438f4aa60fd38d14778" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762449 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-scripts\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762516 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-run-httpd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762553 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-log-httpd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762567 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-config-data\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762587 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762685 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv8kd\" (UniqueName: \"kubernetes.io/projected/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-kube-api-access-qv8kd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762763 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.762932 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.864950 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-log-httpd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.865002 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-config-data\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.865065 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.865484 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-log-httpd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.865787 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv8kd\" (UniqueName: \"kubernetes.io/projected/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-kube-api-access-qv8kd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.865870 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.866017 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.866095 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-scripts\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.866180 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-run-httpd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.866603 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-run-httpd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.869907 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-scripts\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.870079 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.870457 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.871272 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-config-data\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.871557 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.884513 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv8kd\" (UniqueName: \"kubernetes.io/projected/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-kube-api-access-qv8kd\") pod \"ceilometer-0\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " pod="openstack/ceilometer-0" Oct 02 13:09:42 crc kubenswrapper[4710]: I1002 13:09:42.920846 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47447a8c-f16f-4519-8b8a-e83c95cf32bd" path="/var/lib/kubelet/pods/47447a8c-f16f-4519-8b8a-e83c95cf32bd/volumes" Oct 02 13:09:43 crc kubenswrapper[4710]: I1002 13:09:43.038435 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:43 crc kubenswrapper[4710]: W1002 13:09:43.493301 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ddc679_474b_4b91_8e63_ec4b33b2f9b7.slice/crio-9f91fe7adb35a51b26a0655d6f9dbd05dfe62e546e5768c0e24f1db6c0c25da5 WatchSource:0}: Error finding container 9f91fe7adb35a51b26a0655d6f9dbd05dfe62e546e5768c0e24f1db6c0c25da5: Status 404 returned error can't find the container with id 9f91fe7adb35a51b26a0655d6f9dbd05dfe62e546e5768c0e24f1db6c0c25da5 Oct 02 13:09:43 crc kubenswrapper[4710]: I1002 13:09:43.494725 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:43 crc kubenswrapper[4710]: I1002 13:09:43.623989 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerStarted","Data":"9f91fe7adb35a51b26a0655d6f9dbd05dfe62e546e5768c0e24f1db6c0c25da5"} Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.459154 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lmkkc"] Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.461299 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.469418 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lmkkc"] Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.497589 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp8dz\" (UniqueName: \"kubernetes.io/projected/be6c2122-9be7-4414-9e44-8a25e8984d40-kube-api-access-jp8dz\") pod \"nova-api-db-create-lmkkc\" (UID: \"be6c2122-9be7-4414-9e44-8a25e8984d40\") " pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.566220 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-gwdnf"] Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.567786 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.577101 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gwdnf"] Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.599865 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp8dz\" (UniqueName: \"kubernetes.io/projected/be6c2122-9be7-4414-9e44-8a25e8984d40-kube-api-access-jp8dz\") pod \"nova-api-db-create-lmkkc\" (UID: \"be6c2122-9be7-4414-9e44-8a25e8984d40\") " pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.620856 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp8dz\" (UniqueName: \"kubernetes.io/projected/be6c2122-9be7-4414-9e44-8a25e8984d40-kube-api-access-jp8dz\") pod \"nova-api-db-create-lmkkc\" (UID: \"be6c2122-9be7-4414-9e44-8a25e8984d40\") " pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.642652 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerStarted","Data":"0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087"} Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.642700 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerStarted","Data":"8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24"} Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.668038 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xq4sp"] Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.670431 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.682205 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xq4sp"] Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.702052 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjqm\" (UniqueName: \"kubernetes.io/projected/33245146-2264-4fcf-a29b-a0272f6936e9-kube-api-access-4jjqm\") pod \"nova-cell0-db-create-gwdnf\" (UID: \"33245146-2264-4fcf-a29b-a0272f6936e9\") " pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.792677 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.803386 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjqm\" (UniqueName: \"kubernetes.io/projected/33245146-2264-4fcf-a29b-a0272f6936e9-kube-api-access-4jjqm\") pod \"nova-cell0-db-create-gwdnf\" (UID: \"33245146-2264-4fcf-a29b-a0272f6936e9\") " pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.803493 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grmgt\" (UniqueName: \"kubernetes.io/projected/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf-kube-api-access-grmgt\") pod \"nova-cell1-db-create-xq4sp\" (UID: \"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf\") " pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.835517 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjqm\" (UniqueName: \"kubernetes.io/projected/33245146-2264-4fcf-a29b-a0272f6936e9-kube-api-access-4jjqm\") pod \"nova-cell0-db-create-gwdnf\" (UID: \"33245146-2264-4fcf-a29b-a0272f6936e9\") " pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.905297 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grmgt\" (UniqueName: \"kubernetes.io/projected/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf-kube-api-access-grmgt\") pod \"nova-cell1-db-create-xq4sp\" (UID: \"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf\") " pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.926424 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grmgt\" (UniqueName: \"kubernetes.io/projected/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf-kube-api-access-grmgt\") pod \"nova-cell1-db-create-xq4sp\" (UID: \"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf\") " pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:44 crc kubenswrapper[4710]: I1002 13:09:44.993520 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:45 crc kubenswrapper[4710]: I1002 13:09:45.007130 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:45 crc kubenswrapper[4710]: I1002 13:09:45.437357 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lmkkc"] Oct 02 13:09:45 crc kubenswrapper[4710]: W1002 13:09:45.440680 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe6c2122_9be7_4414_9e44_8a25e8984d40.slice/crio-8123fec4ed16ee514c9835aa4028e0186607eb5791764926e51168c631e4ccbe WatchSource:0}: Error finding container 8123fec4ed16ee514c9835aa4028e0186607eb5791764926e51168c631e4ccbe: Status 404 returned error can't find the container with id 8123fec4ed16ee514c9835aa4028e0186607eb5791764926e51168c631e4ccbe Oct 02 13:09:45 crc kubenswrapper[4710]: I1002 13:09:45.660459 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lmkkc" event={"ID":"be6c2122-9be7-4414-9e44-8a25e8984d40","Type":"ContainerStarted","Data":"8123fec4ed16ee514c9835aa4028e0186607eb5791764926e51168c631e4ccbe"} Oct 02 13:09:45 crc kubenswrapper[4710]: I1002 13:09:45.667052 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerStarted","Data":"db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2"} Oct 02 13:09:45 crc kubenswrapper[4710]: I1002 13:09:45.692957 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xq4sp"] Oct 02 13:09:45 crc kubenswrapper[4710]: W1002 13:09:45.726936 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b28f29_2760_43a5_b8fc_af19ad1a3bcf.slice/crio-6504fb69174abeaa75578a643ca850ad6ea8ea03944726f396e2ac160d6ef41f WatchSource:0}: Error finding container 6504fb69174abeaa75578a643ca850ad6ea8ea03944726f396e2ac160d6ef41f: Status 404 returned error can't find the container with id 6504fb69174abeaa75578a643ca850ad6ea8ea03944726f396e2ac160d6ef41f Oct 02 13:09:45 crc kubenswrapper[4710]: I1002 13:09:45.796025 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gwdnf"] Oct 02 13:09:45 crc kubenswrapper[4710]: W1002 13:09:45.804999 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33245146_2264_4fcf_a29b_a0272f6936e9.slice/crio-f102674ee8b0ec33fae5aea7e634fbd2d4cfe31fc78a889e67c6d9a6d0947a48 WatchSource:0}: Error finding container f102674ee8b0ec33fae5aea7e634fbd2d4cfe31fc78a889e67c6d9a6d0947a48: Status 404 returned error can't find the container with id f102674ee8b0ec33fae5aea7e634fbd2d4cfe31fc78a889e67c6d9a6d0947a48 Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.679346 4710 generic.go:334] "Generic (PLEG): container finished" podID="33245146-2264-4fcf-a29b-a0272f6936e9" containerID="05b6f86e9266f0e71fffb41fd202dfcc2bc3f3c463e28376d6127873ce718b37" exitCode=0 Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.680078 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gwdnf" event={"ID":"33245146-2264-4fcf-a29b-a0272f6936e9","Type":"ContainerDied","Data":"05b6f86e9266f0e71fffb41fd202dfcc2bc3f3c463e28376d6127873ce718b37"} Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.680107 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gwdnf" event={"ID":"33245146-2264-4fcf-a29b-a0272f6936e9","Type":"ContainerStarted","Data":"f102674ee8b0ec33fae5aea7e634fbd2d4cfe31fc78a889e67c6d9a6d0947a48"} Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.681941 4710 generic.go:334] "Generic (PLEG): container finished" podID="b8b28f29-2760-43a5-b8fc-af19ad1a3bcf" containerID="c2df07b5e6a8bff7df42122cbea490fef586921805eeb66fbdd30ea523845b96" exitCode=0 Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.681944 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xq4sp" event={"ID":"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf","Type":"ContainerDied","Data":"c2df07b5e6a8bff7df42122cbea490fef586921805eeb66fbdd30ea523845b96"} Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.682108 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xq4sp" event={"ID":"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf","Type":"ContainerStarted","Data":"6504fb69174abeaa75578a643ca850ad6ea8ea03944726f396e2ac160d6ef41f"} Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.685586 4710 generic.go:334] "Generic (PLEG): container finished" podID="be6c2122-9be7-4414-9e44-8a25e8984d40" containerID="3eae176180b2ecda618a4b9bcd0edcad5d1477673b15adef6751adb2d2d9d8fc" exitCode=0 Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.685637 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lmkkc" event={"ID":"be6c2122-9be7-4414-9e44-8a25e8984d40","Type":"ContainerDied","Data":"3eae176180b2ecda618a4b9bcd0edcad5d1477673b15adef6751adb2d2d9d8fc"} Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.948987 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:46 crc kubenswrapper[4710]: I1002 13:09:46.984472 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:47 crc kubenswrapper[4710]: I1002 13:09:47.697011 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerStarted","Data":"a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16"} Oct 02 13:09:47 crc kubenswrapper[4710]: I1002 13:09:47.697440 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:47 crc kubenswrapper[4710]: I1002 13:09:47.730977 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.407948043 podStartE2EDuration="5.730959354s" podCreationTimestamp="2025-10-02 13:09:42 +0000 UTC" firstStartedPulling="2025-10-02 13:09:43.496579759 +0000 UTC m=+1227.602990652" lastFinishedPulling="2025-10-02 13:09:46.81959108 +0000 UTC m=+1230.926001963" observedRunningTime="2025-10-02 13:09:47.725183498 +0000 UTC m=+1231.831594381" watchObservedRunningTime="2025-10-02 13:09:47.730959354 +0000 UTC m=+1231.837370237" Oct 02 13:09:47 crc kubenswrapper[4710]: I1002 13:09:47.736650 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.277341 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.285194 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.291913 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.389583 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp8dz\" (UniqueName: \"kubernetes.io/projected/be6c2122-9be7-4414-9e44-8a25e8984d40-kube-api-access-jp8dz\") pod \"be6c2122-9be7-4414-9e44-8a25e8984d40\" (UID: \"be6c2122-9be7-4414-9e44-8a25e8984d40\") " Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.389632 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grmgt\" (UniqueName: \"kubernetes.io/projected/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf-kube-api-access-grmgt\") pod \"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf\" (UID: \"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf\") " Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.389797 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jjqm\" (UniqueName: \"kubernetes.io/projected/33245146-2264-4fcf-a29b-a0272f6936e9-kube-api-access-4jjqm\") pod \"33245146-2264-4fcf-a29b-a0272f6936e9\" (UID: \"33245146-2264-4fcf-a29b-a0272f6936e9\") " Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.401227 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6c2122-9be7-4414-9e44-8a25e8984d40-kube-api-access-jp8dz" (OuterVolumeSpecName: "kube-api-access-jp8dz") pod "be6c2122-9be7-4414-9e44-8a25e8984d40" (UID: "be6c2122-9be7-4414-9e44-8a25e8984d40"). InnerVolumeSpecName "kube-api-access-jp8dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.402373 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33245146-2264-4fcf-a29b-a0272f6936e9-kube-api-access-4jjqm" (OuterVolumeSpecName: "kube-api-access-4jjqm") pod "33245146-2264-4fcf-a29b-a0272f6936e9" (UID: "33245146-2264-4fcf-a29b-a0272f6936e9"). InnerVolumeSpecName "kube-api-access-4jjqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.404405 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf-kube-api-access-grmgt" (OuterVolumeSpecName: "kube-api-access-grmgt") pod "b8b28f29-2760-43a5-b8fc-af19ad1a3bcf" (UID: "b8b28f29-2760-43a5-b8fc-af19ad1a3bcf"). InnerVolumeSpecName "kube-api-access-grmgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.491759 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp8dz\" (UniqueName: \"kubernetes.io/projected/be6c2122-9be7-4414-9e44-8a25e8984d40-kube-api-access-jp8dz\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.491807 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grmgt\" (UniqueName: \"kubernetes.io/projected/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf-kube-api-access-grmgt\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.491819 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jjqm\" (UniqueName: \"kubernetes.io/projected/33245146-2264-4fcf-a29b-a0272f6936e9-kube-api-access-4jjqm\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.707279 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gwdnf" event={"ID":"33245146-2264-4fcf-a29b-a0272f6936e9","Type":"ContainerDied","Data":"f102674ee8b0ec33fae5aea7e634fbd2d4cfe31fc78a889e67c6d9a6d0947a48"} Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.707591 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f102674ee8b0ec33fae5aea7e634fbd2d4cfe31fc78a889e67c6d9a6d0947a48" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.707306 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gwdnf" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.708900 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xq4sp" event={"ID":"b8b28f29-2760-43a5-b8fc-af19ad1a3bcf","Type":"ContainerDied","Data":"6504fb69174abeaa75578a643ca850ad6ea8ea03944726f396e2ac160d6ef41f"} Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.708922 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6504fb69174abeaa75578a643ca850ad6ea8ea03944726f396e2ac160d6ef41f" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.708932 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xq4sp" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.710509 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lmkkc" event={"ID":"be6c2122-9be7-4414-9e44-8a25e8984d40","Type":"ContainerDied","Data":"8123fec4ed16ee514c9835aa4028e0186607eb5791764926e51168c631e4ccbe"} Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.710542 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8123fec4ed16ee514c9835aa4028e0186607eb5791764926e51168c631e4ccbe" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.710611 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lmkkc" Oct 02 13:09:48 crc kubenswrapper[4710]: I1002 13:09:48.710969 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.433127 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.433946 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-central-agent" containerID="cri-o://8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24" gracePeriod=30 Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.433995 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="sg-core" containerID="cri-o://db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2" gracePeriod=30 Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.434032 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="proxy-httpd" containerID="cri-o://a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16" gracePeriod=30 Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.434043 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-notification-agent" containerID="cri-o://0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087" gracePeriod=30 Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.765624 4710 generic.go:334] "Generic (PLEG): container finished" podID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerID="a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16" exitCode=0 Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.765664 4710 generic.go:334] "Generic (PLEG): container finished" podID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerID="db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2" exitCode=2 Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.765659 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerDied","Data":"a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16"} Oct 02 13:09:53 crc kubenswrapper[4710]: I1002 13:09:53.765703 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerDied","Data":"db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2"} Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.713593 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-eb66-account-create-bcf9b"] Oct 02 13:09:54 crc kubenswrapper[4710]: E1002 13:09:54.714375 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33245146-2264-4fcf-a29b-a0272f6936e9" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.714389 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="33245146-2264-4fcf-a29b-a0272f6936e9" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: E1002 13:09:54.714415 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b28f29-2760-43a5-b8fc-af19ad1a3bcf" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.714421 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b28f29-2760-43a5-b8fc-af19ad1a3bcf" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: E1002 13:09:54.714434 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6c2122-9be7-4414-9e44-8a25e8984d40" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.714440 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6c2122-9be7-4414-9e44-8a25e8984d40" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.714636 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="be6c2122-9be7-4414-9e44-8a25e8984d40" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.714655 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="33245146-2264-4fcf-a29b-a0272f6936e9" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.714665 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b28f29-2760-43a5-b8fc-af19ad1a3bcf" containerName="mariadb-database-create" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.715361 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.717937 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.745631 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-eb66-account-create-bcf9b"] Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.779332 4710 generic.go:334] "Generic (PLEG): container finished" podID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerID="8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24" exitCode=0 Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.779387 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerDied","Data":"8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24"} Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.811486 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7w4h\" (UniqueName: \"kubernetes.io/projected/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47-kube-api-access-r7w4h\") pod \"nova-api-eb66-account-create-bcf9b\" (UID: \"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47\") " pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.889047 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-d881-account-create-nvqlz"] Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.890623 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.893295 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.913506 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7w4h\" (UniqueName: \"kubernetes.io/projected/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47-kube-api-access-r7w4h\") pod \"nova-api-eb66-account-create-bcf9b\" (UID: \"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47\") " pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.914871 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d881-account-create-nvqlz"] Oct 02 13:09:54 crc kubenswrapper[4710]: I1002 13:09:54.935779 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7w4h\" (UniqueName: \"kubernetes.io/projected/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47-kube-api-access-r7w4h\") pod \"nova-api-eb66-account-create-bcf9b\" (UID: \"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47\") " pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.015694 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpttv\" (UniqueName: \"kubernetes.io/projected/1c8a8393-59d4-4c27-9800-d96b2bc96b42-kube-api-access-fpttv\") pod \"nova-cell0-d881-account-create-nvqlz\" (UID: \"1c8a8393-59d4-4c27-9800-d96b2bc96b42\") " pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.043906 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.109028 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-94ae-account-create-5r4pn"] Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.110334 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.116543 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.118716 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpttv\" (UniqueName: \"kubernetes.io/projected/1c8a8393-59d4-4c27-9800-d96b2bc96b42-kube-api-access-fpttv\") pod \"nova-cell0-d881-account-create-nvqlz\" (UID: \"1c8a8393-59d4-4c27-9800-d96b2bc96b42\") " pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.136993 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-94ae-account-create-5r4pn"] Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.148357 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpttv\" (UniqueName: \"kubernetes.io/projected/1c8a8393-59d4-4c27-9800-d96b2bc96b42-kube-api-access-fpttv\") pod \"nova-cell0-d881-account-create-nvqlz\" (UID: \"1c8a8393-59d4-4c27-9800-d96b2bc96b42\") " pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.210709 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.223919 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zqj4\" (UniqueName: \"kubernetes.io/projected/0f9bab2f-dc35-45c4-a7bd-fae3a3576042-kube-api-access-9zqj4\") pod \"nova-cell1-94ae-account-create-5r4pn\" (UID: \"0f9bab2f-dc35-45c4-a7bd-fae3a3576042\") " pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.307486 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.326325 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zqj4\" (UniqueName: \"kubernetes.io/projected/0f9bab2f-dc35-45c4-a7bd-fae3a3576042-kube-api-access-9zqj4\") pod \"nova-cell1-94ae-account-create-5r4pn\" (UID: \"0f9bab2f-dc35-45c4-a7bd-fae3a3576042\") " pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.357577 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zqj4\" (UniqueName: \"kubernetes.io/projected/0f9bab2f-dc35-45c4-a7bd-fae3a3576042-kube-api-access-9zqj4\") pod \"nova-cell1-94ae-account-create-5r4pn\" (UID: \"0f9bab2f-dc35-45c4-a7bd-fae3a3576042\") " pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.427713 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-log-httpd\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428111 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv8kd\" (UniqueName: \"kubernetes.io/projected/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-kube-api-access-qv8kd\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428304 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-config-data\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428340 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-run-httpd\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428369 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-sg-core-conf-yaml\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428411 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-ceilometer-tls-certs\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428435 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428568 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-scripts\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428591 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-combined-ca-bundle\") pod \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\" (UID: \"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7\") " Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.428726 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.429192 4710 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.429213 4710 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.437115 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-scripts" (OuterVolumeSpecName: "scripts") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.437165 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-kube-api-access-qv8kd" (OuterVolumeSpecName: "kube-api-access-qv8kd") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "kube-api-access-qv8kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.467823 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.524204 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.528368 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.533022 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv8kd\" (UniqueName: \"kubernetes.io/projected/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-kube-api-access-qv8kd\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.533057 4710 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.533068 4710 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.533078 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.533088 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.539018 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.581133 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-config-data" (OuterVolumeSpecName: "config-data") pod "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" (UID: "a6ddc679-474b-4b91-8e63-ec4b33b2f9b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.635133 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.636616 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-eb66-account-create-bcf9b"] Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.779407 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d881-account-create-nvqlz"] Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.794786 4710 generic.go:334] "Generic (PLEG): container finished" podID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerID="0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087" exitCode=0 Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.794848 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerDied","Data":"0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087"} Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.794864 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.794890 4710 scope.go:117] "RemoveContainer" containerID="a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.794878 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6ddc679-474b-4b91-8e63-ec4b33b2f9b7","Type":"ContainerDied","Data":"9f91fe7adb35a51b26a0655d6f9dbd05dfe62e546e5768c0e24f1db6c0c25da5"} Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.798958 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eb66-account-create-bcf9b" event={"ID":"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47","Type":"ContainerStarted","Data":"f588b4b4fa59340a542f27f6fa95ee4a43626310e509d4987edae80dc74fbb3b"} Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.854302 4710 scope.go:117] "RemoveContainer" containerID="db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.943672 4710 scope.go:117] "RemoveContainer" containerID="0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.959348 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.972300 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.985032 4710 scope.go:117] "RemoveContainer" containerID="8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989130 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:55 crc kubenswrapper[4710]: E1002 13:09:55.989580 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-notification-agent" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989604 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-notification-agent" Oct 02 13:09:55 crc kubenswrapper[4710]: E1002 13:09:55.989614 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="sg-core" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989621 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="sg-core" Oct 02 13:09:55 crc kubenswrapper[4710]: E1002 13:09:55.989630 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="proxy-httpd" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989637 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="proxy-httpd" Oct 02 13:09:55 crc kubenswrapper[4710]: E1002 13:09:55.989662 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-central-agent" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989668 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-central-agent" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989902 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="sg-core" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989925 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-notification-agent" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989948 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="ceilometer-central-agent" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.989964 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" containerName="proxy-httpd" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.992206 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:55 crc kubenswrapper[4710]: I1002 13:09:55.997208 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.000199 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.000434 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.000538 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.028843 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-94ae-account-create-5r4pn"] Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.038570 4710 scope.go:117] "RemoveContainer" containerID="a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16" Oct 02 13:09:56 crc kubenswrapper[4710]: E1002 13:09:56.039063 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16\": container with ID starting with a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16 not found: ID does not exist" containerID="a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.039108 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16"} err="failed to get container status \"a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16\": rpc error: code = NotFound desc = could not find container \"a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16\": container with ID starting with a1ad1bf55f4075d8a4edd81884ab173dc442aa59daa86da4c38f90b04c720e16 not found: ID does not exist" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.039141 4710 scope.go:117] "RemoveContainer" containerID="db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2" Oct 02 13:09:56 crc kubenswrapper[4710]: E1002 13:09:56.040336 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2\": container with ID starting with db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2 not found: ID does not exist" containerID="db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.040374 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2"} err="failed to get container status \"db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2\": rpc error: code = NotFound desc = could not find container \"db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2\": container with ID starting with db23428c2de58bf8453ad45e84ac3dd72ba35efebb484893f74584adf31915a2 not found: ID does not exist" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.040398 4710 scope.go:117] "RemoveContainer" containerID="0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087" Oct 02 13:09:56 crc kubenswrapper[4710]: E1002 13:09:56.040772 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087\": container with ID starting with 0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087 not found: ID does not exist" containerID="0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.040797 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087"} err="failed to get container status \"0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087\": rpc error: code = NotFound desc = could not find container \"0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087\": container with ID starting with 0d7f800f8ba45cf8f302985d645381dd188b2da6990a7bbe11e272e660402087 not found: ID does not exist" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.040815 4710 scope.go:117] "RemoveContainer" containerID="8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24" Oct 02 13:09:56 crc kubenswrapper[4710]: E1002 13:09:56.041187 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24\": container with ID starting with 8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24 not found: ID does not exist" containerID="8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.041214 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24"} err="failed to get container status \"8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24\": rpc error: code = NotFound desc = could not find container \"8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24\": container with ID starting with 8932d30e449830c014329b4aaeba703bbeac652839371ef58cdb6a6422286b24 not found: ID does not exist" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.145632 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.145924 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-log-httpd\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.145974 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-scripts\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.145994 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6tgm\" (UniqueName: \"kubernetes.io/projected/ab18b551-f6ad-41b5-bd92-b98701675ff6-kube-api-access-x6tgm\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.146016 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-run-httpd\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.146050 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.146073 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.146101 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-config-data\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.250422 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.257709 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.257910 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-log-httpd\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258069 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-scripts\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258119 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6tgm\" (UniqueName: \"kubernetes.io/projected/ab18b551-f6ad-41b5-bd92-b98701675ff6-kube-api-access-x6tgm\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258173 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-run-httpd\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258263 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258305 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-log-httpd\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258321 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.258417 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-config-data\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.259491 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-run-httpd\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.262481 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.262513 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-config-data\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.265938 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-scripts\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.266033 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.278987 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6tgm\" (UniqueName: \"kubernetes.io/projected/ab18b551-f6ad-41b5-bd92-b98701675ff6-kube-api-access-x6tgm\") pod \"ceilometer-0\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.328731 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.796065 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:09:56 crc kubenswrapper[4710]: W1002 13:09:56.803820 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab18b551_f6ad_41b5_bd92_b98701675ff6.slice/crio-43cd86de84ed02da06acc596f78e0f11d5df141b968360c71ede366b352c09d0 WatchSource:0}: Error finding container 43cd86de84ed02da06acc596f78e0f11d5df141b968360c71ede366b352c09d0: Status 404 returned error can't find the container with id 43cd86de84ed02da06acc596f78e0f11d5df141b968360c71ede366b352c09d0 Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.823946 4710 generic.go:334] "Generic (PLEG): container finished" podID="0f9bab2f-dc35-45c4-a7bd-fae3a3576042" containerID="cf79824db6fe9cd2b13ea05cc96c32838c6226bfeb29d1e1acf00214650fe939" exitCode=0 Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.824033 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-94ae-account-create-5r4pn" event={"ID":"0f9bab2f-dc35-45c4-a7bd-fae3a3576042","Type":"ContainerDied","Data":"cf79824db6fe9cd2b13ea05cc96c32838c6226bfeb29d1e1acf00214650fe939"} Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.824065 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-94ae-account-create-5r4pn" event={"ID":"0f9bab2f-dc35-45c4-a7bd-fae3a3576042","Type":"ContainerStarted","Data":"52a331580962cc39452c84345a9675207af46d43d0e12acd1ad8a8879a81c086"} Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.828567 4710 generic.go:334] "Generic (PLEG): container finished" podID="63bbbc57-c9bc-43fd-8bd9-15845ff6cd47" containerID="347b532dea3695b2c69836fe68848317a286ff6c35e596e623921f1058c51b2c" exitCode=0 Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.828671 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eb66-account-create-bcf9b" event={"ID":"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47","Type":"ContainerDied","Data":"347b532dea3695b2c69836fe68848317a286ff6c35e596e623921f1058c51b2c"} Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.830959 4710 generic.go:334] "Generic (PLEG): container finished" podID="1c8a8393-59d4-4c27-9800-d96b2bc96b42" containerID="0add1d91184851a1fef685997472e3d576979615052d37cefd689ec6c2d31e52" exitCode=0 Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.831014 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d881-account-create-nvqlz" event={"ID":"1c8a8393-59d4-4c27-9800-d96b2bc96b42","Type":"ContainerDied","Data":"0add1d91184851a1fef685997472e3d576979615052d37cefd689ec6c2d31e52"} Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.831044 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d881-account-create-nvqlz" event={"ID":"1c8a8393-59d4-4c27-9800-d96b2bc96b42","Type":"ContainerStarted","Data":"5a4ce3d56fa0c2aa01c97e663ce527986466de73e32686989d2dfcb153350b0a"} Oct 02 13:09:56 crc kubenswrapper[4710]: I1002 13:09:56.964183 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ddc679-474b-4b91-8e63-ec4b33b2f9b7" path="/var/lib/kubelet/pods/a6ddc679-474b-4b91-8e63-ec4b33b2f9b7/volumes" Oct 02 13:09:57 crc kubenswrapper[4710]: I1002 13:09:57.844681 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerStarted","Data":"a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba"} Oct 02 13:09:57 crc kubenswrapper[4710]: I1002 13:09:57.845329 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerStarted","Data":"cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651"} Oct 02 13:09:57 crc kubenswrapper[4710]: I1002 13:09:57.845343 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerStarted","Data":"43cd86de84ed02da06acc596f78e0f11d5df141b968360c71ede366b352c09d0"} Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.434592 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.445559 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.452796 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.534775 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7w4h\" (UniqueName: \"kubernetes.io/projected/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47-kube-api-access-r7w4h\") pod \"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47\" (UID: \"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47\") " Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.540638 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47-kube-api-access-r7w4h" (OuterVolumeSpecName: "kube-api-access-r7w4h") pod "63bbbc57-c9bc-43fd-8bd9-15845ff6cd47" (UID: "63bbbc57-c9bc-43fd-8bd9-15845ff6cd47"). InnerVolumeSpecName "kube-api-access-r7w4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.636409 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpttv\" (UniqueName: \"kubernetes.io/projected/1c8a8393-59d4-4c27-9800-d96b2bc96b42-kube-api-access-fpttv\") pod \"1c8a8393-59d4-4c27-9800-d96b2bc96b42\" (UID: \"1c8a8393-59d4-4c27-9800-d96b2bc96b42\") " Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.636702 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zqj4\" (UniqueName: \"kubernetes.io/projected/0f9bab2f-dc35-45c4-a7bd-fae3a3576042-kube-api-access-9zqj4\") pod \"0f9bab2f-dc35-45c4-a7bd-fae3a3576042\" (UID: \"0f9bab2f-dc35-45c4-a7bd-fae3a3576042\") " Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.637332 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7w4h\" (UniqueName: \"kubernetes.io/projected/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47-kube-api-access-r7w4h\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.640650 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c8a8393-59d4-4c27-9800-d96b2bc96b42-kube-api-access-fpttv" (OuterVolumeSpecName: "kube-api-access-fpttv") pod "1c8a8393-59d4-4c27-9800-d96b2bc96b42" (UID: "1c8a8393-59d4-4c27-9800-d96b2bc96b42"). InnerVolumeSpecName "kube-api-access-fpttv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.640919 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f9bab2f-dc35-45c4-a7bd-fae3a3576042-kube-api-access-9zqj4" (OuterVolumeSpecName: "kube-api-access-9zqj4") pod "0f9bab2f-dc35-45c4-a7bd-fae3a3576042" (UID: "0f9bab2f-dc35-45c4-a7bd-fae3a3576042"). InnerVolumeSpecName "kube-api-access-9zqj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.739488 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpttv\" (UniqueName: \"kubernetes.io/projected/1c8a8393-59d4-4c27-9800-d96b2bc96b42-kube-api-access-fpttv\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.739543 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zqj4\" (UniqueName: \"kubernetes.io/projected/0f9bab2f-dc35-45c4-a7bd-fae3a3576042-kube-api-access-9zqj4\") on node \"crc\" DevicePath \"\"" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.860541 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eb66-account-create-bcf9b" event={"ID":"63bbbc57-c9bc-43fd-8bd9-15845ff6cd47","Type":"ContainerDied","Data":"f588b4b4fa59340a542f27f6fa95ee4a43626310e509d4987edae80dc74fbb3b"} Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.860581 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f588b4b4fa59340a542f27f6fa95ee4a43626310e509d4987edae80dc74fbb3b" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.860556 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eb66-account-create-bcf9b" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.864884 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d881-account-create-nvqlz" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.864881 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d881-account-create-nvqlz" event={"ID":"1c8a8393-59d4-4c27-9800-d96b2bc96b42","Type":"ContainerDied","Data":"5a4ce3d56fa0c2aa01c97e663ce527986466de73e32686989d2dfcb153350b0a"} Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.864999 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a4ce3d56fa0c2aa01c97e663ce527986466de73e32686989d2dfcb153350b0a" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.866922 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerStarted","Data":"c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6"} Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.868824 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-94ae-account-create-5r4pn" event={"ID":"0f9bab2f-dc35-45c4-a7bd-fae3a3576042","Type":"ContainerDied","Data":"52a331580962cc39452c84345a9675207af46d43d0e12acd1ad8a8879a81c086"} Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.868852 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52a331580962cc39452c84345a9675207af46d43d0e12acd1ad8a8879a81c086" Oct 02 13:09:58 crc kubenswrapper[4710]: I1002 13:09:58.868862 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ae-account-create-5r4pn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.146285 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l6gmn"] Oct 02 13:10:00 crc kubenswrapper[4710]: E1002 13:10:00.151616 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bbbc57-c9bc-43fd-8bd9-15845ff6cd47" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.151653 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bbbc57-c9bc-43fd-8bd9-15845ff6cd47" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: E1002 13:10:00.151666 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c8a8393-59d4-4c27-9800-d96b2bc96b42" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.151672 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c8a8393-59d4-4c27-9800-d96b2bc96b42" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: E1002 13:10:00.151703 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9bab2f-dc35-45c4-a7bd-fae3a3576042" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.151708 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9bab2f-dc35-45c4-a7bd-fae3a3576042" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.151922 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="63bbbc57-c9bc-43fd-8bd9-15845ff6cd47" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.151933 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f9bab2f-dc35-45c4-a7bd-fae3a3576042" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.151940 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c8a8393-59d4-4c27-9800-d96b2bc96b42" containerName="mariadb-account-create" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.152582 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.156186 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.156393 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b95hm" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.156624 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.162694 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l6gmn"] Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.172831 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-config-data\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.172873 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-scripts\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.172894 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jptsd\" (UniqueName: \"kubernetes.io/projected/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-kube-api-access-jptsd\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.172919 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.277009 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-config-data\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.277066 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-scripts\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.277096 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jptsd\" (UniqueName: \"kubernetes.io/projected/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-kube-api-access-jptsd\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.277131 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.284447 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.284477 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-scripts\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.292560 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-config-data\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.297260 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jptsd\" (UniqueName: \"kubernetes.io/projected/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-kube-api-access-jptsd\") pod \"nova-cell0-conductor-db-sync-l6gmn\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.475170 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.893288 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerStarted","Data":"32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e"} Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.893876 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.933037 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.860470137 podStartE2EDuration="5.933017812s" podCreationTimestamp="2025-10-02 13:09:55 +0000 UTC" firstStartedPulling="2025-10-02 13:09:56.807625237 +0000 UTC m=+1240.914036120" lastFinishedPulling="2025-10-02 13:09:59.880172902 +0000 UTC m=+1243.986583795" observedRunningTime="2025-10-02 13:10:00.929699828 +0000 UTC m=+1245.036110711" watchObservedRunningTime="2025-10-02 13:10:00.933017812 +0000 UTC m=+1245.039428695" Oct 02 13:10:00 crc kubenswrapper[4710]: I1002 13:10:00.973318 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l6gmn"] Oct 02 13:10:01 crc kubenswrapper[4710]: I1002 13:10:01.914000 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" event={"ID":"ff3004d0-bcfc-490a-bc66-0b14469fb0a5","Type":"ContainerStarted","Data":"3a20670913ccb193f15aded310ba4d724d7201f85fa7c4a3f7a2be085954a84e"} Oct 02 13:10:10 crc kubenswrapper[4710]: I1002 13:10:10.998211 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" event={"ID":"ff3004d0-bcfc-490a-bc66-0b14469fb0a5","Type":"ContainerStarted","Data":"63cd89d4d3711584f92f2146a20d5616ce5357f11db77e97d9dbff50ba00aa3a"} Oct 02 13:10:11 crc kubenswrapper[4710]: I1002 13:10:11.037392 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" podStartSLOduration=2.027409949 podStartE2EDuration="11.037373497s" podCreationTimestamp="2025-10-02 13:10:00 +0000 UTC" firstStartedPulling="2025-10-02 13:10:00.986959553 +0000 UTC m=+1245.093370436" lastFinishedPulling="2025-10-02 13:10:09.996923101 +0000 UTC m=+1254.103333984" observedRunningTime="2025-10-02 13:10:11.02808041 +0000 UTC m=+1255.134491293" watchObservedRunningTime="2025-10-02 13:10:11.037373497 +0000 UTC m=+1255.143784380" Oct 02 13:10:26 crc kubenswrapper[4710]: I1002 13:10:26.349143 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 13:10:34 crc kubenswrapper[4710]: I1002 13:10:34.209955 4710 generic.go:334] "Generic (PLEG): container finished" podID="ff3004d0-bcfc-490a-bc66-0b14469fb0a5" containerID="63cd89d4d3711584f92f2146a20d5616ce5357f11db77e97d9dbff50ba00aa3a" exitCode=0 Oct 02 13:10:34 crc kubenswrapper[4710]: I1002 13:10:34.210934 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" event={"ID":"ff3004d0-bcfc-490a-bc66-0b14469fb0a5","Type":"ContainerDied","Data":"63cd89d4d3711584f92f2146a20d5616ce5357f11db77e97d9dbff50ba00aa3a"} Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.566718 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.630627 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-scripts\") pod \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.630797 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jptsd\" (UniqueName: \"kubernetes.io/projected/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-kube-api-access-jptsd\") pod \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.631082 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-combined-ca-bundle\") pod \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.636876 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-scripts" (OuterVolumeSpecName: "scripts") pod "ff3004d0-bcfc-490a-bc66-0b14469fb0a5" (UID: "ff3004d0-bcfc-490a-bc66-0b14469fb0a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.648034 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-kube-api-access-jptsd" (OuterVolumeSpecName: "kube-api-access-jptsd") pod "ff3004d0-bcfc-490a-bc66-0b14469fb0a5" (UID: "ff3004d0-bcfc-490a-bc66-0b14469fb0a5"). InnerVolumeSpecName "kube-api-access-jptsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.660217 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff3004d0-bcfc-490a-bc66-0b14469fb0a5" (UID: "ff3004d0-bcfc-490a-bc66-0b14469fb0a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.731962 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-config-data\") pod \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\" (UID: \"ff3004d0-bcfc-490a-bc66-0b14469fb0a5\") " Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.732340 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.732353 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jptsd\" (UniqueName: \"kubernetes.io/projected/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-kube-api-access-jptsd\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.732364 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.760113 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-config-data" (OuterVolumeSpecName: "config-data") pod "ff3004d0-bcfc-490a-bc66-0b14469fb0a5" (UID: "ff3004d0-bcfc-490a-bc66-0b14469fb0a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:10:35 crc kubenswrapper[4710]: I1002 13:10:35.834023 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3004d0-bcfc-490a-bc66-0b14469fb0a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.231094 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" event={"ID":"ff3004d0-bcfc-490a-bc66-0b14469fb0a5","Type":"ContainerDied","Data":"3a20670913ccb193f15aded310ba4d724d7201f85fa7c4a3f7a2be085954a84e"} Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.231384 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a20670913ccb193f15aded310ba4d724d7201f85fa7c4a3f7a2be085954a84e" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.231337 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l6gmn" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.321399 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 13:10:36 crc kubenswrapper[4710]: E1002 13:10:36.321874 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3004d0-bcfc-490a-bc66-0b14469fb0a5" containerName="nova-cell0-conductor-db-sync" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.321904 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3004d0-bcfc-490a-bc66-0b14469fb0a5" containerName="nova-cell0-conductor-db-sync" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.322138 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3004d0-bcfc-490a-bc66-0b14469fb0a5" containerName="nova-cell0-conductor-db-sync" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.322996 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.326487 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b95hm" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.326727 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.331627 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.444627 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.445285 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ckrv\" (UniqueName: \"kubernetes.io/projected/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-kube-api-access-5ckrv\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.445645 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.548168 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ckrv\" (UniqueName: \"kubernetes.io/projected/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-kube-api-access-5ckrv\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.548472 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.548621 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.554334 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.554371 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.576542 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ckrv\" (UniqueName: \"kubernetes.io/projected/c3ab8e0c-8d4b-41b2-af0b-24ebf616425e-kube-api-access-5ckrv\") pod \"nova-cell0-conductor-0\" (UID: \"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:36 crc kubenswrapper[4710]: I1002 13:10:36.641898 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:37 crc kubenswrapper[4710]: I1002 13:10:37.135154 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 13:10:37 crc kubenswrapper[4710]: I1002 13:10:37.246428 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e","Type":"ContainerStarted","Data":"d5fef8b741ceab824caad3b1abe60ee1875a463564057dd2d36dda2110083f6b"} Oct 02 13:10:38 crc kubenswrapper[4710]: I1002 13:10:38.257097 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c3ab8e0c-8d4b-41b2-af0b-24ebf616425e","Type":"ContainerStarted","Data":"82eaf08000675eb14f6a3a75893a4dc10a2a410974e991df8546d2ce0e2c4308"} Oct 02 13:10:38 crc kubenswrapper[4710]: I1002 13:10:38.257605 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:46 crc kubenswrapper[4710]: I1002 13:10:46.684538 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 13:10:46 crc kubenswrapper[4710]: I1002 13:10:46.711579 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=10.71154547 podStartE2EDuration="10.71154547s" podCreationTimestamp="2025-10-02 13:10:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:10:38.278312212 +0000 UTC m=+1282.384723095" watchObservedRunningTime="2025-10-02 13:10:46.71154547 +0000 UTC m=+1290.817956353" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.163275 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fn4fw"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.164776 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.169297 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.169369 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.174568 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fn4fw"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.261661 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-scripts\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.261857 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-config-data\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.261945 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8rr2\" (UniqueName: \"kubernetes.io/projected/32dda3a2-0573-4560-8303-8edc0838c7b9-kube-api-access-f8rr2\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.262094 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.329419 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.331738 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.334031 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.343698 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.365634 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8rr2\" (UniqueName: \"kubernetes.io/projected/32dda3a2-0573-4560-8303-8edc0838c7b9-kube-api-access-f8rr2\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.365732 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.365878 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-scripts\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.365966 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-config-data\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.390489 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-scripts\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.390513 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-config-data\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.409379 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8rr2\" (UniqueName: \"kubernetes.io/projected/32dda3a2-0573-4560-8303-8edc0838c7b9-kube-api-access-f8rr2\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.410437 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fn4fw\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.447263 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.449575 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.464696 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.473624 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.483026 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-config-data\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.483332 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-logs\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.483473 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.483617 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h985j\" (UniqueName: \"kubernetes.io/projected/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-kube-api-access-h985j\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.502195 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.539871 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.541790 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.543982 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.562826 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.591726 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-config-data\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.606934 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-logs\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.607017 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.607081 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.607162 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h985j\" (UniqueName: \"kubernetes.io/projected/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-kube-api-access-h985j\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.607183 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g92gl\" (UniqueName: \"kubernetes.io/projected/0dbd175b-387d-4f12-823a-2830e899db24-kube-api-access-g92gl\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.607323 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dbd175b-387d-4f12-823a-2830e899db24-logs\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.607466 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-config-data\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.608804 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-logs\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.629905 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.639828 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-config-data\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.655878 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h985j\" (UniqueName: \"kubernetes.io/projected/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-kube-api-access-h985j\") pod \"nova-api-0\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.656420 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.710062 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.714515 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715196 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dbd175b-387d-4f12-823a-2830e899db24-logs\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715281 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-config-data\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715313 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715341 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715377 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltdss\" (UniqueName: \"kubernetes.io/projected/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-kube-api-access-ltdss\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715409 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g92gl\" (UniqueName: \"kubernetes.io/projected/0dbd175b-387d-4f12-823a-2830e899db24-kube-api-access-g92gl\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.715426 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-config-data\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.716096 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dbd175b-387d-4f12-823a-2830e899db24-logs\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.721620 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-config-data\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.725844 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.738998 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b455c8df9-g6bzr"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.741585 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.753533 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.768160 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.774149 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g92gl\" (UniqueName: \"kubernetes.io/projected/0dbd175b-387d-4f12-823a-2830e899db24-kube-api-access-g92gl\") pod \"nova-metadata-0\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.785517 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b455c8df9-g6bzr"] Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.817532 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.817633 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.817760 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltdss\" (UniqueName: \"kubernetes.io/projected/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-kube-api-access-ltdss\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.817798 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.817840 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-config-data\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.817939 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlql6\" (UniqueName: \"kubernetes.io/projected/3a8608c4-2369-4b74-927a-00eea2f1ca6e-kube-api-access-jlql6\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.826068 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.832504 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-config-data\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.844373 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltdss\" (UniqueName: \"kubernetes.io/projected/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-kube-api-access-ltdss\") pod \"nova-scheduler-0\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " pod="openstack/nova-scheduler-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.859269 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.921098 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lptfl\" (UniqueName: \"kubernetes.io/projected/237922f1-0c3a-4d6f-9404-2a9fe3efa393-kube-api-access-lptfl\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.921440 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-svc\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.922112 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlql6\" (UniqueName: \"kubernetes.io/projected/3a8608c4-2369-4b74-927a-00eea2f1ca6e-kube-api-access-jlql6\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.922185 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-swift-storage-0\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.922223 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-nb\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.922594 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-config\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.922653 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.922959 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.923169 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-sb\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.927662 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.930098 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:47 crc kubenswrapper[4710]: I1002 13:10:47.945692 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlql6\" (UniqueName: \"kubernetes.io/projected/3a8608c4-2369-4b74-927a-00eea2f1ca6e-kube-api-access-jlql6\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.025366 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-svc\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.025455 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-swift-storage-0\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.025481 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-nb\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.025539 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-config\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.025716 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-sb\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.026855 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-config\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.027416 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-swift-storage-0\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.027635 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-nb\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.027966 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-svc\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.028135 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lptfl\" (UniqueName: \"kubernetes.io/projected/237922f1-0c3a-4d6f-9404-2a9fe3efa393-kube-api-access-lptfl\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.028648 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-sb\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.058638 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lptfl\" (UniqueName: \"kubernetes.io/projected/237922f1-0c3a-4d6f-9404-2a9fe3efa393-kube-api-access-lptfl\") pod \"dnsmasq-dns-b455c8df9-g6bzr\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.073249 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.097179 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.112087 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.274887 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fn4fw"] Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.293370 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.330451 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djnzg"] Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.331891 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.342887 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.342939 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.381891 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djnzg"] Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.414988 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.462520 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-config-data\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.462712 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.463088 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwdq\" (UniqueName: \"kubernetes.io/projected/21d98a36-dd9d-42b5-b26c-a6fa30189c44-kube-api-access-fmwdq\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.463169 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-scripts\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.498523 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fn4fw" event={"ID":"32dda3a2-0573-4560-8303-8edc0838c7b9","Type":"ContainerStarted","Data":"07746b7e920d6a369e57d8149e69804d5a821d2ac06259c7d88741b1d3a81d6a"} Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.537944 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dbd175b-387d-4f12-823a-2830e899db24","Type":"ContainerStarted","Data":"6524a9cf3c0ac5c5618c4ac4efb9abc4128b58916921cbd539303918441fc867"} Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.566428 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-config-data\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.566493 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.566574 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwdq\" (UniqueName: \"kubernetes.io/projected/21d98a36-dd9d-42b5-b26c-a6fa30189c44-kube-api-access-fmwdq\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.566598 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-scripts\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.591180 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.598646 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-scripts\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.606481 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-config-data\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.638554 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwdq\" (UniqueName: \"kubernetes.io/projected/21d98a36-dd9d-42b5-b26c-a6fa30189c44-kube-api-access-fmwdq\") pod \"nova-cell1-conductor-db-sync-djnzg\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.775237 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:10:48 crc kubenswrapper[4710]: I1002 13:10:48.854397 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.156136 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.169060 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b455c8df9-g6bzr"] Oct 02 13:10:49 crc kubenswrapper[4710]: W1002 13:10:49.182533 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod237922f1_0c3a_4d6f_9404_2a9fe3efa393.slice/crio-1b612905e9a1f55fc10aabee6482a868e4f03797b0fbc0a9bbf0cbe377aae9dd WatchSource:0}: Error finding container 1b612905e9a1f55fc10aabee6482a868e4f03797b0fbc0a9bbf0cbe377aae9dd: Status 404 returned error can't find the container with id 1b612905e9a1f55fc10aabee6482a868e4f03797b0fbc0a9bbf0cbe377aae9dd Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.505566 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djnzg"] Oct 02 13:10:49 crc kubenswrapper[4710]: W1002 13:10:49.530935 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21d98a36_dd9d_42b5_b26c_a6fa30189c44.slice/crio-44820c2000f4dd7d09a6e3d40453dd83aafd2ceddcf7b9cd903a4f25e37b7f10 WatchSource:0}: Error finding container 44820c2000f4dd7d09a6e3d40453dd83aafd2ceddcf7b9cd903a4f25e37b7f10: Status 404 returned error can't find the container with id 44820c2000f4dd7d09a6e3d40453dd83aafd2ceddcf7b9cd903a4f25e37b7f10 Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.572423 4710 generic.go:334] "Generic (PLEG): container finished" podID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerID="332a5602d52bf3b1a2204ab29137ee19e9f8f589a06a4a6540e7af05212ab2d8" exitCode=0 Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.572533 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" event={"ID":"237922f1-0c3a-4d6f-9404-2a9fe3efa393","Type":"ContainerDied","Data":"332a5602d52bf3b1a2204ab29137ee19e9f8f589a06a4a6540e7af05212ab2d8"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.572564 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" event={"ID":"237922f1-0c3a-4d6f-9404-2a9fe3efa393","Type":"ContainerStarted","Data":"1b612905e9a1f55fc10aabee6482a868e4f03797b0fbc0a9bbf0cbe377aae9dd"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.580881 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fn4fw" event={"ID":"32dda3a2-0573-4560-8303-8edc0838c7b9","Type":"ContainerStarted","Data":"c1dc86d76c795049c88be02c4b18e3c9223d8331f1fd017439982d2b75df666b"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.591659 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6","Type":"ContainerStarted","Data":"42a0d2aa28d79bd467a1fe18685ccfa3e5b1e313214458947e474bad13ca6596"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.606413 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djnzg" event={"ID":"21d98a36-dd9d-42b5-b26c-a6fa30189c44","Type":"ContainerStarted","Data":"44820c2000f4dd7d09a6e3d40453dd83aafd2ceddcf7b9cd903a4f25e37b7f10"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.624003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990","Type":"ContainerStarted","Data":"908ec24650987b682ba96309f93779d678f7df69c43d3aa47c35c193cddf9a11"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.626022 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a8608c4-2369-4b74-927a-00eea2f1ca6e","Type":"ContainerStarted","Data":"181a54f63b2b177c3a27f3b2cf4108572ce00d1007f1cf3c3948afbda784efba"} Oct 02 13:10:49 crc kubenswrapper[4710]: I1002 13:10:49.679758 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fn4fw" podStartSLOduration=2.679715178 podStartE2EDuration="2.679715178s" podCreationTimestamp="2025-10-02 13:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:10:49.623419217 +0000 UTC m=+1293.729830110" watchObservedRunningTime="2025-10-02 13:10:49.679715178 +0000 UTC m=+1293.786126061" Oct 02 13:10:50 crc kubenswrapper[4710]: I1002 13:10:50.653250 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djnzg" event={"ID":"21d98a36-dd9d-42b5-b26c-a6fa30189c44","Type":"ContainerStarted","Data":"0160887fdcb5d2cfdf4ca31b2bdcebc3b9ad92ad0f7c4375a93527f91ee161a9"} Oct 02 13:10:50 crc kubenswrapper[4710]: I1002 13:10:50.658251 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" event={"ID":"237922f1-0c3a-4d6f-9404-2a9fe3efa393","Type":"ContainerStarted","Data":"64db8b97a25ab0c2628ad9533e43803a084b5f3ae23dabfc0942ae00ce911995"} Oct 02 13:10:50 crc kubenswrapper[4710]: I1002 13:10:50.676423 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-djnzg" podStartSLOduration=2.676393837 podStartE2EDuration="2.676393837s" podCreationTimestamp="2025-10-02 13:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:10:50.670875668 +0000 UTC m=+1294.777286551" watchObservedRunningTime="2025-10-02 13:10:50.676393837 +0000 UTC m=+1294.782804720" Oct 02 13:10:50 crc kubenswrapper[4710]: I1002 13:10:50.694823 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" podStartSLOduration=3.694800002 podStartE2EDuration="3.694800002s" podCreationTimestamp="2025-10-02 13:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:10:50.689776445 +0000 UTC m=+1294.796187328" watchObservedRunningTime="2025-10-02 13:10:50.694800002 +0000 UTC m=+1294.801210895" Oct 02 13:10:51 crc kubenswrapper[4710]: I1002 13:10:51.398239 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:51 crc kubenswrapper[4710]: I1002 13:10:51.410025 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:10:51 crc kubenswrapper[4710]: I1002 13:10:51.669913 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.697992 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990","Type":"ContainerStarted","Data":"a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8"} Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.698557 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990","Type":"ContainerStarted","Data":"02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1"} Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.702018 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a8608c4-2369-4b74-927a-00eea2f1ca6e","Type":"ContainerStarted","Data":"0c5aba13999b99e49da83eb41df108889038dff1ec92b039f7873ea54061e1b6"} Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.702156 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3a8608c4-2369-4b74-927a-00eea2f1ca6e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0c5aba13999b99e49da83eb41df108889038dff1ec92b039f7873ea54061e1b6" gracePeriod=30 Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.706788 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6","Type":"ContainerStarted","Data":"9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16"} Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.711472 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dbd175b-387d-4f12-823a-2830e899db24","Type":"ContainerStarted","Data":"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a"} Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.711526 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dbd175b-387d-4f12-823a-2830e899db24","Type":"ContainerStarted","Data":"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1"} Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.711680 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-log" containerID="cri-o://c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1" gracePeriod=30 Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.712018 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-metadata" containerID="cri-o://7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a" gracePeriod=30 Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.752701 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.883090512 podStartE2EDuration="7.752675737s" podCreationTimestamp="2025-10-02 13:10:47 +0000 UTC" firstStartedPulling="2025-10-02 13:10:48.787867925 +0000 UTC m=+1292.894278808" lastFinishedPulling="2025-10-02 13:10:53.65745315 +0000 UTC m=+1297.763864033" observedRunningTime="2025-10-02 13:10:54.749707282 +0000 UTC m=+1298.856118165" watchObservedRunningTime="2025-10-02 13:10:54.752675737 +0000 UTC m=+1298.859086620" Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.757160 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.552952871 podStartE2EDuration="7.75714572s" podCreationTimestamp="2025-10-02 13:10:47 +0000 UTC" firstStartedPulling="2025-10-02 13:10:48.463156821 +0000 UTC m=+1292.569567704" lastFinishedPulling="2025-10-02 13:10:53.66734967 +0000 UTC m=+1297.773760553" observedRunningTime="2025-10-02 13:10:54.724370992 +0000 UTC m=+1298.830781875" watchObservedRunningTime="2025-10-02 13:10:54.75714572 +0000 UTC m=+1298.863556603" Oct 02 13:10:54 crc kubenswrapper[4710]: I1002 13:10:54.773556 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.530195969 podStartE2EDuration="7.773539204s" podCreationTimestamp="2025-10-02 13:10:47 +0000 UTC" firstStartedPulling="2025-10-02 13:10:48.414349141 +0000 UTC m=+1292.520760024" lastFinishedPulling="2025-10-02 13:10:53.657692376 +0000 UTC m=+1297.764103259" observedRunningTime="2025-10-02 13:10:54.766481466 +0000 UTC m=+1298.872892349" watchObservedRunningTime="2025-10-02 13:10:54.773539204 +0000 UTC m=+1298.879950077" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.364349 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.384851 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.910477371 podStartE2EDuration="8.384834441s" podCreationTimestamp="2025-10-02 13:10:47 +0000 UTC" firstStartedPulling="2025-10-02 13:10:49.182946336 +0000 UTC m=+1293.289357219" lastFinishedPulling="2025-10-02 13:10:53.657303406 +0000 UTC m=+1297.763714289" observedRunningTime="2025-10-02 13:10:54.786525202 +0000 UTC m=+1298.892936085" watchObservedRunningTime="2025-10-02 13:10:55.384834441 +0000 UTC m=+1299.491245324" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.445681 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-config-data\") pod \"0dbd175b-387d-4f12-823a-2830e899db24\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.445725 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g92gl\" (UniqueName: \"kubernetes.io/projected/0dbd175b-387d-4f12-823a-2830e899db24-kube-api-access-g92gl\") pod \"0dbd175b-387d-4f12-823a-2830e899db24\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.445769 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dbd175b-387d-4f12-823a-2830e899db24-logs\") pod \"0dbd175b-387d-4f12-823a-2830e899db24\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.445911 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-combined-ca-bundle\") pod \"0dbd175b-387d-4f12-823a-2830e899db24\" (UID: \"0dbd175b-387d-4f12-823a-2830e899db24\") " Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.446135 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbd175b-387d-4f12-823a-2830e899db24-logs" (OuterVolumeSpecName: "logs") pod "0dbd175b-387d-4f12-823a-2830e899db24" (UID: "0dbd175b-387d-4f12-823a-2830e899db24"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.446466 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dbd175b-387d-4f12-823a-2830e899db24-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.452483 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dbd175b-387d-4f12-823a-2830e899db24-kube-api-access-g92gl" (OuterVolumeSpecName: "kube-api-access-g92gl") pod "0dbd175b-387d-4f12-823a-2830e899db24" (UID: "0dbd175b-387d-4f12-823a-2830e899db24"). InnerVolumeSpecName "kube-api-access-g92gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.478094 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dbd175b-387d-4f12-823a-2830e899db24" (UID: "0dbd175b-387d-4f12-823a-2830e899db24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.488557 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-config-data" (OuterVolumeSpecName: "config-data") pod "0dbd175b-387d-4f12-823a-2830e899db24" (UID: "0dbd175b-387d-4f12-823a-2830e899db24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.549775 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.549872 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g92gl\" (UniqueName: \"kubernetes.io/projected/0dbd175b-387d-4f12-823a-2830e899db24-kube-api-access-g92gl\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.549889 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbd175b-387d-4f12-823a-2830e899db24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.722261 4710 generic.go:334] "Generic (PLEG): container finished" podID="0dbd175b-387d-4f12-823a-2830e899db24" containerID="7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a" exitCode=0 Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.722293 4710 generic.go:334] "Generic (PLEG): container finished" podID="0dbd175b-387d-4f12-823a-2830e899db24" containerID="c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1" exitCode=143 Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.723202 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.723443 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dbd175b-387d-4f12-823a-2830e899db24","Type":"ContainerDied","Data":"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a"} Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.723477 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dbd175b-387d-4f12-823a-2830e899db24","Type":"ContainerDied","Data":"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1"} Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.723491 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dbd175b-387d-4f12-823a-2830e899db24","Type":"ContainerDied","Data":"6524a9cf3c0ac5c5618c4ac4efb9abc4128b58916921cbd539303918441fc867"} Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.723509 4710 scope.go:117] "RemoveContainer" containerID="7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.772831 4710 scope.go:117] "RemoveContainer" containerID="c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.782781 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.796771 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.798627 4710 scope.go:117] "RemoveContainer" containerID="7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a" Oct 02 13:10:55 crc kubenswrapper[4710]: E1002 13:10:55.799393 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a\": container with ID starting with 7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a not found: ID does not exist" containerID="7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.799447 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a"} err="failed to get container status \"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a\": rpc error: code = NotFound desc = could not find container \"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a\": container with ID starting with 7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a not found: ID does not exist" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.799476 4710 scope.go:117] "RemoveContainer" containerID="c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1" Oct 02 13:10:55 crc kubenswrapper[4710]: E1002 13:10:55.800169 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1\": container with ID starting with c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1 not found: ID does not exist" containerID="c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.800207 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1"} err="failed to get container status \"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1\": rpc error: code = NotFound desc = could not find container \"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1\": container with ID starting with c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1 not found: ID does not exist" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.800226 4710 scope.go:117] "RemoveContainer" containerID="7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.800507 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a"} err="failed to get container status \"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a\": rpc error: code = NotFound desc = could not find container \"7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a\": container with ID starting with 7f4b753b0a4a4e951ce3241d0447c0234feb5fe43f9f2e2e238aeddf48461d1a not found: ID does not exist" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.800538 4710 scope.go:117] "RemoveContainer" containerID="c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.800846 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1"} err="failed to get container status \"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1\": rpc error: code = NotFound desc = could not find container \"c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1\": container with ID starting with c87d86113d9884d8cc5e02788d1f2322ad726631a37108ff11e180044d5103b1 not found: ID does not exist" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.810202 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:55 crc kubenswrapper[4710]: E1002 13:10:55.811520 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-metadata" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.811602 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-metadata" Oct 02 13:10:55 crc kubenswrapper[4710]: E1002 13:10:55.811685 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-log" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.811696 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-log" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.812107 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-metadata" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.812142 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbd175b-387d-4f12-823a-2830e899db24" containerName="nova-metadata-log" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.813644 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.816947 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.816947 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.830032 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.861182 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.861304 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4smxd\" (UniqueName: \"kubernetes.io/projected/d1d637ca-08fe-4819-8766-c0e93aea8d73-kube-api-access-4smxd\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.861420 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d637ca-08fe-4819-8766-c0e93aea8d73-logs\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.861522 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-config-data\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.861559 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.964069 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d637ca-08fe-4819-8766-c0e93aea8d73-logs\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.964455 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-config-data\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.964487 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.964515 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d637ca-08fe-4819-8766-c0e93aea8d73-logs\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.965997 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.966593 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4smxd\" (UniqueName: \"kubernetes.io/projected/d1d637ca-08fe-4819-8766-c0e93aea8d73-kube-api-access-4smxd\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.970024 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.970048 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-config-data\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.983112 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:55 crc kubenswrapper[4710]: I1002 13:10:55.984421 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4smxd\" (UniqueName: \"kubernetes.io/projected/d1d637ca-08fe-4819-8766-c0e93aea8d73-kube-api-access-4smxd\") pod \"nova-metadata-0\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " pod="openstack/nova-metadata-0" Oct 02 13:10:56 crc kubenswrapper[4710]: I1002 13:10:56.138429 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:10:56 crc kubenswrapper[4710]: I1002 13:10:56.687017 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:10:56 crc kubenswrapper[4710]: I1002 13:10:56.736274 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d1d637ca-08fe-4819-8766-c0e93aea8d73","Type":"ContainerStarted","Data":"1c976a745cd996e43be94ccb83fcdaee0e96dbc867793e7598f4101efc7a7f87"} Oct 02 13:10:56 crc kubenswrapper[4710]: I1002 13:10:56.916255 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dbd175b-387d-4f12-823a-2830e899db24" path="/var/lib/kubelet/pods/0dbd175b-387d-4f12-823a-2830e899db24/volumes" Oct 02 13:10:57 crc kubenswrapper[4710]: I1002 13:10:57.658863 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:10:57 crc kubenswrapper[4710]: I1002 13:10:57.659209 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:10:57 crc kubenswrapper[4710]: I1002 13:10:57.751546 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d1d637ca-08fe-4819-8766-c0e93aea8d73","Type":"ContainerStarted","Data":"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341"} Oct 02 13:10:57 crc kubenswrapper[4710]: I1002 13:10:57.751591 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d1d637ca-08fe-4819-8766-c0e93aea8d73","Type":"ContainerStarted","Data":"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085"} Oct 02 13:10:57 crc kubenswrapper[4710]: I1002 13:10:57.786783 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.786734227 podStartE2EDuration="2.786734227s" podCreationTimestamp="2025-10-02 13:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:10:57.769563024 +0000 UTC m=+1301.875973917" watchObservedRunningTime="2025-10-02 13:10:57.786734227 +0000 UTC m=+1301.893145110" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.075172 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.075474 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.098124 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.112815 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.113877 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.202317 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9bc865865-4f875"] Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.203015 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9bc865865-4f875" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerName="dnsmasq-dns" containerID="cri-o://d599285d2de11783bf93fa6dc9c6be10e9a0fe21db62e996d8ec62843903b065" gracePeriod=10 Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.741048 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.741053 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.767033 4710 generic.go:334] "Generic (PLEG): container finished" podID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerID="d599285d2de11783bf93fa6dc9c6be10e9a0fe21db62e996d8ec62843903b065" exitCode=0 Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.768694 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bc865865-4f875" event={"ID":"335e25f6-16e6-4aaa-a1b0-32595ab2591e","Type":"ContainerDied","Data":"d599285d2de11783bf93fa6dc9c6be10e9a0fe21db62e996d8ec62843903b065"} Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.768773 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9bc865865-4f875" event={"ID":"335e25f6-16e6-4aaa-a1b0-32595ab2591e","Type":"ContainerDied","Data":"d8d3dbd93dace3af58024d5126b46a224e9a5e8c8c85a6561bd9819cf748d4ec"} Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.768787 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8d3dbd93dace3af58024d5126b46a224e9a5e8c8c85a6561bd9819cf748d4ec" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.771965 4710 generic.go:334] "Generic (PLEG): container finished" podID="32dda3a2-0573-4560-8303-8edc0838c7b9" containerID="c1dc86d76c795049c88be02c4b18e3c9223d8331f1fd017439982d2b75df666b" exitCode=0 Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.773695 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fn4fw" event={"ID":"32dda3a2-0573-4560-8303-8edc0838c7b9","Type":"ContainerDied","Data":"c1dc86d76c795049c88be02c4b18e3c9223d8331f1fd017439982d2b75df666b"} Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.809150 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 13:10:58 crc kubenswrapper[4710]: I1002 13:10:58.939076 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.032566 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-sb\") pod \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.033753 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ldgp\" (UniqueName: \"kubernetes.io/projected/335e25f6-16e6-4aaa-a1b0-32595ab2591e-kube-api-access-8ldgp\") pod \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.033786 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-config\") pod \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.033900 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-swift-storage-0\") pod \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.033922 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-svc\") pod \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.034033 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-nb\") pod \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\" (UID: \"335e25f6-16e6-4aaa-a1b0-32595ab2591e\") " Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.039550 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/335e25f6-16e6-4aaa-a1b0-32595ab2591e-kube-api-access-8ldgp" (OuterVolumeSpecName: "kube-api-access-8ldgp") pod "335e25f6-16e6-4aaa-a1b0-32595ab2591e" (UID: "335e25f6-16e6-4aaa-a1b0-32595ab2591e"). InnerVolumeSpecName "kube-api-access-8ldgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.088415 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "335e25f6-16e6-4aaa-a1b0-32595ab2591e" (UID: "335e25f6-16e6-4aaa-a1b0-32595ab2591e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.099222 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "335e25f6-16e6-4aaa-a1b0-32595ab2591e" (UID: "335e25f6-16e6-4aaa-a1b0-32595ab2591e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.107874 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "335e25f6-16e6-4aaa-a1b0-32595ab2591e" (UID: "335e25f6-16e6-4aaa-a1b0-32595ab2591e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.110568 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "335e25f6-16e6-4aaa-a1b0-32595ab2591e" (UID: "335e25f6-16e6-4aaa-a1b0-32595ab2591e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.111683 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-config" (OuterVolumeSpecName: "config") pod "335e25f6-16e6-4aaa-a1b0-32595ab2591e" (UID: "335e25f6-16e6-4aaa-a1b0-32595ab2591e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.136461 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.136505 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.136522 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ldgp\" (UniqueName: \"kubernetes.io/projected/335e25f6-16e6-4aaa-a1b0-32595ab2591e-kube-api-access-8ldgp\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.136537 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.136550 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.136561 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/335e25f6-16e6-4aaa-a1b0-32595ab2591e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.783697 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9bc865865-4f875" Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.827916 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9bc865865-4f875"] Oct 02 13:10:59 crc kubenswrapper[4710]: I1002 13:10:59.845789 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9bc865865-4f875"] Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.205279 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.358951 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-combined-ca-bundle\") pod \"32dda3a2-0573-4560-8303-8edc0838c7b9\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.359416 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8rr2\" (UniqueName: \"kubernetes.io/projected/32dda3a2-0573-4560-8303-8edc0838c7b9-kube-api-access-f8rr2\") pod \"32dda3a2-0573-4560-8303-8edc0838c7b9\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.359486 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-scripts\") pod \"32dda3a2-0573-4560-8303-8edc0838c7b9\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.359524 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-config-data\") pod \"32dda3a2-0573-4560-8303-8edc0838c7b9\" (UID: \"32dda3a2-0573-4560-8303-8edc0838c7b9\") " Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.378817 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-scripts" (OuterVolumeSpecName: "scripts") pod "32dda3a2-0573-4560-8303-8edc0838c7b9" (UID: "32dda3a2-0573-4560-8303-8edc0838c7b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.383362 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32dda3a2-0573-4560-8303-8edc0838c7b9-kube-api-access-f8rr2" (OuterVolumeSpecName: "kube-api-access-f8rr2") pod "32dda3a2-0573-4560-8303-8edc0838c7b9" (UID: "32dda3a2-0573-4560-8303-8edc0838c7b9"). InnerVolumeSpecName "kube-api-access-f8rr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.396659 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-config-data" (OuterVolumeSpecName: "config-data") pod "32dda3a2-0573-4560-8303-8edc0838c7b9" (UID: "32dda3a2-0573-4560-8303-8edc0838c7b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.397175 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32dda3a2-0573-4560-8303-8edc0838c7b9" (UID: "32dda3a2-0573-4560-8303-8edc0838c7b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.462517 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8rr2\" (UniqueName: \"kubernetes.io/projected/32dda3a2-0573-4560-8303-8edc0838c7b9-kube-api-access-f8rr2\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.462561 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.462575 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.462588 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32dda3a2-0573-4560-8303-8edc0838c7b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.797417 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fn4fw" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.810143 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fn4fw" event={"ID":"32dda3a2-0573-4560-8303-8edc0838c7b9","Type":"ContainerDied","Data":"07746b7e920d6a369e57d8149e69804d5a821d2ac06259c7d88741b1d3a81d6a"} Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.810196 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07746b7e920d6a369e57d8149e69804d5a821d2ac06259c7d88741b1d3a81d6a" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.929152 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" path="/var/lib/kubelet/pods/335e25f6-16e6-4aaa-a1b0-32595ab2591e/volumes" Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.985434 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.985784 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-log" containerID="cri-o://02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1" gracePeriod=30 Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.985854 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-api" containerID="cri-o://a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8" gracePeriod=30 Oct 02 13:11:00 crc kubenswrapper[4710]: I1002 13:11:00.996356 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.081917 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.082481 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-log" containerID="cri-o://b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085" gracePeriod=30 Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.082874 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-metadata" containerID="cri-o://866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341" gracePeriod=30 Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.138635 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.138728 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.664719 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.789424 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-combined-ca-bundle\") pod \"d1d637ca-08fe-4819-8766-c0e93aea8d73\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.789473 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-config-data\") pod \"d1d637ca-08fe-4819-8766-c0e93aea8d73\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.789499 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-nova-metadata-tls-certs\") pod \"d1d637ca-08fe-4819-8766-c0e93aea8d73\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.789606 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4smxd\" (UniqueName: \"kubernetes.io/projected/d1d637ca-08fe-4819-8766-c0e93aea8d73-kube-api-access-4smxd\") pod \"d1d637ca-08fe-4819-8766-c0e93aea8d73\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.789693 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d637ca-08fe-4819-8766-c0e93aea8d73-logs\") pod \"d1d637ca-08fe-4819-8766-c0e93aea8d73\" (UID: \"d1d637ca-08fe-4819-8766-c0e93aea8d73\") " Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.790631 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d637ca-08fe-4819-8766-c0e93aea8d73-logs" (OuterVolumeSpecName: "logs") pod "d1d637ca-08fe-4819-8766-c0e93aea8d73" (UID: "d1d637ca-08fe-4819-8766-c0e93aea8d73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.795689 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d637ca-08fe-4819-8766-c0e93aea8d73-kube-api-access-4smxd" (OuterVolumeSpecName: "kube-api-access-4smxd") pod "d1d637ca-08fe-4819-8766-c0e93aea8d73" (UID: "d1d637ca-08fe-4819-8766-c0e93aea8d73"). InnerVolumeSpecName "kube-api-access-4smxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810507 4710 generic.go:334] "Generic (PLEG): container finished" podID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerID="866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341" exitCode=0 Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810543 4710 generic.go:334] "Generic (PLEG): container finished" podID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerID="b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085" exitCode=143 Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810556 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810566 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d1d637ca-08fe-4819-8766-c0e93aea8d73","Type":"ContainerDied","Data":"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341"} Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810640 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d1d637ca-08fe-4819-8766-c0e93aea8d73","Type":"ContainerDied","Data":"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085"} Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810656 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d1d637ca-08fe-4819-8766-c0e93aea8d73","Type":"ContainerDied","Data":"1c976a745cd996e43be94ccb83fcdaee0e96dbc867793e7598f4101efc7a7f87"} Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.810712 4710 scope.go:117] "RemoveContainer" containerID="866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.816555 4710 generic.go:334] "Generic (PLEG): container finished" podID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerID="02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1" exitCode=143 Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.816600 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990","Type":"ContainerDied","Data":"02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1"} Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.816933 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" containerName="nova-scheduler-scheduler" containerID="cri-o://9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16" gracePeriod=30 Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.829655 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1d637ca-08fe-4819-8766-c0e93aea8d73" (UID: "d1d637ca-08fe-4819-8766-c0e93aea8d73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.837000 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-config-data" (OuterVolumeSpecName: "config-data") pod "d1d637ca-08fe-4819-8766-c0e93aea8d73" (UID: "d1d637ca-08fe-4819-8766-c0e93aea8d73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.857499 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d1d637ca-08fe-4819-8766-c0e93aea8d73" (UID: "d1d637ca-08fe-4819-8766-c0e93aea8d73"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.891719 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.891775 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.891788 4710 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d637ca-08fe-4819-8766-c0e93aea8d73-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.891801 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4smxd\" (UniqueName: \"kubernetes.io/projected/d1d637ca-08fe-4819-8766-c0e93aea8d73-kube-api-access-4smxd\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.891811 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d637ca-08fe-4819-8766-c0e93aea8d73-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.956411 4710 scope.go:117] "RemoveContainer" containerID="b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.975183 4710 scope.go:117] "RemoveContainer" containerID="866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341" Oct 02 13:11:01 crc kubenswrapper[4710]: E1002 13:11:01.975653 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341\": container with ID starting with 866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341 not found: ID does not exist" containerID="866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.975693 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341"} err="failed to get container status \"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341\": rpc error: code = NotFound desc = could not find container \"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341\": container with ID starting with 866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341 not found: ID does not exist" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.975727 4710 scope.go:117] "RemoveContainer" containerID="b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085" Oct 02 13:11:01 crc kubenswrapper[4710]: E1002 13:11:01.976214 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085\": container with ID starting with b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085 not found: ID does not exist" containerID="b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.976240 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085"} err="failed to get container status \"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085\": rpc error: code = NotFound desc = could not find container \"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085\": container with ID starting with b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085 not found: ID does not exist" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.976256 4710 scope.go:117] "RemoveContainer" containerID="866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.976468 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341"} err="failed to get container status \"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341\": rpc error: code = NotFound desc = could not find container \"866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341\": container with ID starting with 866c7a5d1f752080ef8c63b9daeb290a0e28e7979cc24ad3515d998aec048341 not found: ID does not exist" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.976547 4710 scope.go:117] "RemoveContainer" containerID="b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085" Oct 02 13:11:01 crc kubenswrapper[4710]: I1002 13:11:01.976858 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085"} err="failed to get container status \"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085\": rpc error: code = NotFound desc = could not find container \"b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085\": container with ID starting with b70a72ed424ae96c307525a6aa3b775a391f47ef18c7fba018e4e3bcc5a9d085 not found: ID does not exist" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.144109 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.152148 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.165476 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:02 crc kubenswrapper[4710]: E1002 13:11:02.166293 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerName="dnsmasq-dns" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.166428 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerName="dnsmasq-dns" Oct 02 13:11:02 crc kubenswrapper[4710]: E1002 13:11:02.166511 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-log" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.166597 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-log" Oct 02 13:11:02 crc kubenswrapper[4710]: E1002 13:11:02.166686 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dda3a2-0573-4560-8303-8edc0838c7b9" containerName="nova-manage" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.166782 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dda3a2-0573-4560-8303-8edc0838c7b9" containerName="nova-manage" Oct 02 13:11:02 crc kubenswrapper[4710]: E1002 13:11:02.166894 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-metadata" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.166973 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-metadata" Oct 02 13:11:02 crc kubenswrapper[4710]: E1002 13:11:02.167061 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerName="init" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.167134 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerName="init" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.167465 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-metadata" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.167568 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" containerName="nova-metadata-log" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.167658 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="32dda3a2-0573-4560-8303-8edc0838c7b9" containerName="nova-manage" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.167772 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="335e25f6-16e6-4aaa-a1b0-32595ab2591e" containerName="dnsmasq-dns" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.169255 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.171855 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.173066 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.179948 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.299513 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.299847 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-logs\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.300003 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-config-data\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.300115 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgnmw\" (UniqueName: \"kubernetes.io/projected/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-kube-api-access-kgnmw\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.300231 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.401858 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.401929 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-logs\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.402005 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-config-data\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.402036 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgnmw\" (UniqueName: \"kubernetes.io/projected/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-kube-api-access-kgnmw\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.402054 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.402672 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-logs\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.407560 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.407589 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.407687 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-config-data\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.421359 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgnmw\" (UniqueName: \"kubernetes.io/projected/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-kube-api-access-kgnmw\") pod \"nova-metadata-0\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.490546 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.837906 4710 generic.go:334] "Generic (PLEG): container finished" podID="21d98a36-dd9d-42b5-b26c-a6fa30189c44" containerID="0160887fdcb5d2cfdf4ca31b2bdcebc3b9ad92ad0f7c4375a93527f91ee161a9" exitCode=0 Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.838271 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djnzg" event={"ID":"21d98a36-dd9d-42b5-b26c-a6fa30189c44","Type":"ContainerDied","Data":"0160887fdcb5d2cfdf4ca31b2bdcebc3b9ad92ad0f7c4375a93527f91ee161a9"} Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.849737 4710 generic.go:334] "Generic (PLEG): container finished" podID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" containerID="9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16" exitCode=0 Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.849793 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6","Type":"ContainerDied","Data":"9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16"} Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.916062 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d637ca-08fe-4819-8766-c0e93aea8d73" path="/var/lib/kubelet/pods/d1d637ca-08fe-4819-8766-c0e93aea8d73/volumes" Oct 02 13:11:02 crc kubenswrapper[4710]: W1002 13:11:02.984970 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44fd49c6_fa11_4c3a_9a95_72ae18059ac7.slice/crio-8bb9774bfe20fe64651fa783617b3fc8c7e691be108f10175ee75669999da20c WatchSource:0}: Error finding container 8bb9774bfe20fe64651fa783617b3fc8c7e691be108f10175ee75669999da20c: Status 404 returned error can't find the container with id 8bb9774bfe20fe64651fa783617b3fc8c7e691be108f10175ee75669999da20c Oct 02 13:11:02 crc kubenswrapper[4710]: I1002 13:11:02.997359 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.091568 4710 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16 is running failed: container process not found" containerID="9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.092044 4710 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16 is running failed: container process not found" containerID="9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.092487 4710 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16 is running failed: container process not found" containerID="9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.092648 4710 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" containerName="nova-scheduler-scheduler" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.233556 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.427053 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-config-data\") pod \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.427358 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltdss\" (UniqueName: \"kubernetes.io/projected/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-kube-api-access-ltdss\") pod \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.427433 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-combined-ca-bundle\") pod \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\" (UID: \"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.436693 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-kube-api-access-ltdss" (OuterVolumeSpecName: "kube-api-access-ltdss") pod "6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" (UID: "6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6"). InnerVolumeSpecName "kube-api-access-ltdss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.466572 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" (UID: "6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.476542 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-config-data" (OuterVolumeSpecName: "config-data") pod "6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" (UID: "6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.530196 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltdss\" (UniqueName: \"kubernetes.io/projected/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-kube-api-access-ltdss\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.530588 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.530609 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.623919 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.733856 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-combined-ca-bundle\") pod \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.734077 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-config-data\") pod \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.734128 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h985j\" (UniqueName: \"kubernetes.io/projected/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-kube-api-access-h985j\") pod \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.734170 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-logs\") pod \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\" (UID: \"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990\") " Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.734926 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-logs" (OuterVolumeSpecName: "logs") pod "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" (UID: "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.737676 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-kube-api-access-h985j" (OuterVolumeSpecName: "kube-api-access-h985j") pod "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" (UID: "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990"). InnerVolumeSpecName "kube-api-access-h985j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.764107 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-config-data" (OuterVolumeSpecName: "config-data") pod "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" (UID: "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.765686 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" (UID: "9d3fc1e4-4eb0-4414-9fc3-0f7314baf990"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.836235 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h985j\" (UniqueName: \"kubernetes.io/projected/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-kube-api-access-h985j\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.836276 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.836286 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.836295 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.861585 4710 generic.go:334] "Generic (PLEG): container finished" podID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerID="a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8" exitCode=0 Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.861634 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.861676 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990","Type":"ContainerDied","Data":"a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8"} Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.861707 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9d3fc1e4-4eb0-4414-9fc3-0f7314baf990","Type":"ContainerDied","Data":"908ec24650987b682ba96309f93779d678f7df69c43d3aa47c35c193cddf9a11"} Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.861727 4710 scope.go:117] "RemoveContainer" containerID="a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.866483 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44fd49c6-fa11-4c3a-9a95-72ae18059ac7","Type":"ContainerStarted","Data":"400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a"} Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.866524 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44fd49c6-fa11-4c3a-9a95-72ae18059ac7","Type":"ContainerStarted","Data":"e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c"} Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.866537 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44fd49c6-fa11-4c3a-9a95-72ae18059ac7","Type":"ContainerStarted","Data":"8bb9774bfe20fe64651fa783617b3fc8c7e691be108f10175ee75669999da20c"} Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.868098 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6","Type":"ContainerDied","Data":"42a0d2aa28d79bd467a1fe18685ccfa3e5b1e313214458947e474bad13ca6596"} Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.868142 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.892842 4710 scope.go:117] "RemoveContainer" containerID="02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.904783 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.904755998 podStartE2EDuration="1.904755998s" podCreationTimestamp="2025-10-02 13:11:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:03.887292337 +0000 UTC m=+1307.993703220" watchObservedRunningTime="2025-10-02 13:11:03.904755998 +0000 UTC m=+1308.011166881" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.921000 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.932515 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.935715 4710 scope.go:117] "RemoveContainer" containerID="a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8" Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.936432 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8\": container with ID starting with a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8 not found: ID does not exist" containerID="a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.936468 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8"} err="failed to get container status \"a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8\": rpc error: code = NotFound desc = could not find container \"a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8\": container with ID starting with a5ee5f9d0581568d029d82c4be18aecaaf1a29c68f01fa9a286d7655e0b3ccd8 not found: ID does not exist" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.936495 4710 scope.go:117] "RemoveContainer" containerID="02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1" Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.937883 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1\": container with ID starting with 02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1 not found: ID does not exist" containerID="02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.937982 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1"} err="failed to get container status \"02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1\": rpc error: code = NotFound desc = could not find container \"02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1\": container with ID starting with 02c737a52d477f8f8eb3544a2de7b7aaadbaacd0245f6bce85931daa66eb11f1 not found: ID does not exist" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.938066 4710 scope.go:117] "RemoveContainer" containerID="9a2ffe4fbc8b1795d5ec0dbcc6e7fe9a5bfcd14165fda36b8e13bbd6cbc11c16" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.942069 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.942661 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" containerName="nova-scheduler-scheduler" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.942684 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" containerName="nova-scheduler-scheduler" Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.942721 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-api" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.942731 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-api" Oct 02 13:11:03 crc kubenswrapper[4710]: E1002 13:11:03.942781 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-log" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.942789 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-log" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.943033 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-log" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.943053 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" containerName="nova-api-api" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.943067 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" containerName="nova-scheduler-scheduler" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.946505 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.949408 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.955071 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.969253 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.986996 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:03 crc kubenswrapper[4710]: I1002 13:11:03.999640 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.010641 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.013355 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.029371 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.141951 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-config-data\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.142033 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-config-data\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.142107 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86952\" (UniqueName: \"kubernetes.io/projected/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-kube-api-access-86952\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.142204 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.142307 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mw89\" (UniqueName: \"kubernetes.io/projected/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-kube-api-access-2mw89\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.142380 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-logs\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.142546 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.244227 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-config-data\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.244305 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-config-data\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.244377 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86952\" (UniqueName: \"kubernetes.io/projected/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-kube-api-access-86952\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.244470 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.244503 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mw89\" (UniqueName: \"kubernetes.io/projected/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-kube-api-access-2mw89\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.244586 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-logs\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.245019 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-logs\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.245120 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.249460 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.249543 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.250064 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-config-data\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.264257 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-config-data\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.265120 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86952\" (UniqueName: \"kubernetes.io/projected/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-kube-api-access-86952\") pod \"nova-scheduler-0\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.266325 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mw89\" (UniqueName: \"kubernetes.io/projected/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-kube-api-access-2mw89\") pod \"nova-api-0\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.283796 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.321516 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.336339 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.448846 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-combined-ca-bundle\") pod \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.449174 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-scripts\") pod \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.449255 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-config-data\") pod \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.449348 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmwdq\" (UniqueName: \"kubernetes.io/projected/21d98a36-dd9d-42b5-b26c-a6fa30189c44-kube-api-access-fmwdq\") pod \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\" (UID: \"21d98a36-dd9d-42b5-b26c-a6fa30189c44\") " Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.455027 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-scripts" (OuterVolumeSpecName: "scripts") pod "21d98a36-dd9d-42b5-b26c-a6fa30189c44" (UID: "21d98a36-dd9d-42b5-b26c-a6fa30189c44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.455178 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d98a36-dd9d-42b5-b26c-a6fa30189c44-kube-api-access-fmwdq" (OuterVolumeSpecName: "kube-api-access-fmwdq") pod "21d98a36-dd9d-42b5-b26c-a6fa30189c44" (UID: "21d98a36-dd9d-42b5-b26c-a6fa30189c44"). InnerVolumeSpecName "kube-api-access-fmwdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.480257 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-config-data" (OuterVolumeSpecName: "config-data") pod "21d98a36-dd9d-42b5-b26c-a6fa30189c44" (UID: "21d98a36-dd9d-42b5-b26c-a6fa30189c44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.482882 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21d98a36-dd9d-42b5-b26c-a6fa30189c44" (UID: "21d98a36-dd9d-42b5-b26c-a6fa30189c44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.552682 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.552725 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.552766 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmwdq\" (UniqueName: \"kubernetes.io/projected/21d98a36-dd9d-42b5-b26c-a6fa30189c44-kube-api-access-fmwdq\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.552788 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d98a36-dd9d-42b5-b26c-a6fa30189c44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.739138 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.831639 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:04 crc kubenswrapper[4710]: W1002 13:11:04.837647 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28bb35ad_6443_4ff0_86f0_358b7fe8ecc3.slice/crio-a671b8be9c39cc294daa5efff68bdf38a8f7d4d158940596a7d7e28c96c21990 WatchSource:0}: Error finding container a671b8be9c39cc294daa5efff68bdf38a8f7d4d158940596a7d7e28c96c21990: Status 404 returned error can't find the container with id a671b8be9c39cc294daa5efff68bdf38a8f7d4d158940596a7d7e28c96c21990 Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.882614 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4eba802-f7d9-4994-89ee-c0d4ab52ab51","Type":"ContainerStarted","Data":"328c6a527d08438877ac01f0598724bd8d298f8d1fdf28ee185ea3f290c04f55"} Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.895856 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3","Type":"ContainerStarted","Data":"a671b8be9c39cc294daa5efff68bdf38a8f7d4d158940596a7d7e28c96c21990"} Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.901654 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djnzg" event={"ID":"21d98a36-dd9d-42b5-b26c-a6fa30189c44","Type":"ContainerDied","Data":"44820c2000f4dd7d09a6e3d40453dd83aafd2ceddcf7b9cd903a4f25e37b7f10"} Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.901706 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44820c2000f4dd7d09a6e3d40453dd83aafd2ceddcf7b9cd903a4f25e37b7f10" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.901823 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.918671 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6" path="/var/lib/kubelet/pods/6fb283a6-f1ae-4c79-94f4-b8f33c23e7f6/volumes" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.919571 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3fc1e4-4eb0-4414-9fc3-0f7314baf990" path="/var/lib/kubelet/pods/9d3fc1e4-4eb0-4414-9fc3-0f7314baf990/volumes" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.950895 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 13:11:04 crc kubenswrapper[4710]: E1002 13:11:04.951663 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d98a36-dd9d-42b5-b26c-a6fa30189c44" containerName="nova-cell1-conductor-db-sync" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.951833 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d98a36-dd9d-42b5-b26c-a6fa30189c44" containerName="nova-cell1-conductor-db-sync" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.952174 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d98a36-dd9d-42b5-b26c-a6fa30189c44" containerName="nova-cell1-conductor-db-sync" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.953116 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.964387 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 13:11:04 crc kubenswrapper[4710]: I1002 13:11:04.965573 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.096396 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f91fa-4bac-4673-a8d3-77135bc34f08-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.101180 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vkjg\" (UniqueName: \"kubernetes.io/projected/339f91fa-4bac-4673-a8d3-77135bc34f08-kube-api-access-9vkjg\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.101305 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f91fa-4bac-4673-a8d3-77135bc34f08-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.203237 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f91fa-4bac-4673-a8d3-77135bc34f08-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.203372 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vkjg\" (UniqueName: \"kubernetes.io/projected/339f91fa-4bac-4673-a8d3-77135bc34f08-kube-api-access-9vkjg\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.203420 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f91fa-4bac-4673-a8d3-77135bc34f08-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.208174 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f91fa-4bac-4673-a8d3-77135bc34f08-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.208986 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f91fa-4bac-4673-a8d3-77135bc34f08-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.224639 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vkjg\" (UniqueName: \"kubernetes.io/projected/339f91fa-4bac-4673-a8d3-77135bc34f08-kube-api-access-9vkjg\") pod \"nova-cell1-conductor-0\" (UID: \"339f91fa-4bac-4673-a8d3-77135bc34f08\") " pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:05 crc kubenswrapper[4710]: I1002 13:11:05.299593 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.750940 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 13:11:06 crc kubenswrapper[4710]: W1002 13:11:05.753130 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod339f91fa_4bac_4673_a8d3_77135bc34f08.slice/crio-79cd1bae8bc391eee0d86367c7dfd752ee5a510980ab40c818d50d928520dddf WatchSource:0}: Error finding container 79cd1bae8bc391eee0d86367c7dfd752ee5a510980ab40c818d50d928520dddf: Status 404 returned error can't find the container with id 79cd1bae8bc391eee0d86367c7dfd752ee5a510980ab40c818d50d928520dddf Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.926797 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4eba802-f7d9-4994-89ee-c0d4ab52ab51","Type":"ContainerStarted","Data":"221bb79a48952b1ae9932896c64fe61d69901d4b080df376ffb34d1129fa647e"} Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.926844 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4eba802-f7d9-4994-89ee-c0d4ab52ab51","Type":"ContainerStarted","Data":"5cfb918a1b80d50a17cb2bf2e97f5fafcdeebf4d9ad359a4449fcd4a9d0fbf30"} Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.932218 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"339f91fa-4bac-4673-a8d3-77135bc34f08","Type":"ContainerStarted","Data":"79cd1bae8bc391eee0d86367c7dfd752ee5a510980ab40c818d50d928520dddf"} Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.935757 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3","Type":"ContainerStarted","Data":"4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9"} Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.947444 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.947424432 podStartE2EDuration="2.947424432s" podCreationTimestamp="2025-10-02 13:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:05.945621297 +0000 UTC m=+1310.052032180" watchObservedRunningTime="2025-10-02 13:11:05.947424432 +0000 UTC m=+1310.053835315" Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:05.969340 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.969301775 podStartE2EDuration="2.969301775s" podCreationTimestamp="2025-10-02 13:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:05.961949049 +0000 UTC m=+1310.068359932" watchObservedRunningTime="2025-10-02 13:11:05.969301775 +0000 UTC m=+1310.075712658" Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:06.947611 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"339f91fa-4bac-4673-a8d3-77135bc34f08","Type":"ContainerStarted","Data":"78d76889364e4ccf882c0d42b3121e4484178bd5d1025382b17a787fab6d9ff0"} Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:06.948198 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:06 crc kubenswrapper[4710]: I1002 13:11:06.967184 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.967166064 podStartE2EDuration="2.967166064s" podCreationTimestamp="2025-10-02 13:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:06.964287852 +0000 UTC m=+1311.070698755" watchObservedRunningTime="2025-10-02 13:11:06.967166064 +0000 UTC m=+1311.073576947" Oct 02 13:11:07 crc kubenswrapper[4710]: I1002 13:11:07.491495 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 13:11:07 crc kubenswrapper[4710]: I1002 13:11:07.491583 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 13:11:09 crc kubenswrapper[4710]: I1002 13:11:09.337532 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 13:11:12 crc kubenswrapper[4710]: I1002 13:11:12.491649 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 13:11:12 crc kubenswrapper[4710]: I1002 13:11:12.492681 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 13:11:13 crc kubenswrapper[4710]: I1002 13:11:13.503888 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:13 crc kubenswrapper[4710]: I1002 13:11:13.503888 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:14 crc kubenswrapper[4710]: I1002 13:11:14.284302 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:11:14 crc kubenswrapper[4710]: I1002 13:11:14.284389 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:11:14 crc kubenswrapper[4710]: I1002 13:11:14.336763 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 13:11:14 crc kubenswrapper[4710]: I1002 13:11:14.365467 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 13:11:15 crc kubenswrapper[4710]: I1002 13:11:15.055904 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 13:11:15 crc kubenswrapper[4710]: I1002 13:11:15.331796 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 13:11:15 crc kubenswrapper[4710]: I1002 13:11:15.369426 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.215:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:15 crc kubenswrapper[4710]: I1002 13:11:15.369813 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.215:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:22 crc kubenswrapper[4710]: I1002 13:11:22.500302 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 13:11:22 crc kubenswrapper[4710]: I1002 13:11:22.502431 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 13:11:22 crc kubenswrapper[4710]: I1002 13:11:22.512966 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 13:11:23 crc kubenswrapper[4710]: I1002 13:11:23.107417 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.292894 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.293333 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.293721 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.293782 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.300739 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.308494 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.495116 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b87b8bb9-9pdwf"] Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.497595 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.510215 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b87b8bb9-9pdwf"] Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.512314 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-svc\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.512365 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xdd6\" (UniqueName: \"kubernetes.io/projected/93d26107-27ce-422c-b17d-ae46b170f8b2-kube-api-access-6xdd6\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.512404 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-swift-storage-0\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.512448 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-config\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.512502 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-nb\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.512600 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-sb\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.614539 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-sb\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.614614 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-svc\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.614678 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xdd6\" (UniqueName: \"kubernetes.io/projected/93d26107-27ce-422c-b17d-ae46b170f8b2-kube-api-access-6xdd6\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.614722 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-swift-storage-0\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.614798 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-config\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.614853 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-nb\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.616064 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-config\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.616299 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-swift-storage-0\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.616528 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-sb\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.616601 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-nb\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.617235 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-svc\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.658142 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xdd6\" (UniqueName: \"kubernetes.io/projected/93d26107-27ce-422c-b17d-ae46b170f8b2-kube-api-access-6xdd6\") pod \"dnsmasq-dns-7b87b8bb9-9pdwf\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:24 crc kubenswrapper[4710]: I1002 13:11:24.837816 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.174836 4710 generic.go:334] "Generic (PLEG): container finished" podID="3a8608c4-2369-4b74-927a-00eea2f1ca6e" containerID="0c5aba13999b99e49da83eb41df108889038dff1ec92b039f7873ea54061e1b6" exitCode=137 Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.174930 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a8608c4-2369-4b74-927a-00eea2f1ca6e","Type":"ContainerDied","Data":"0c5aba13999b99e49da83eb41df108889038dff1ec92b039f7873ea54061e1b6"} Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.306367 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.329807 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlql6\" (UniqueName: \"kubernetes.io/projected/3a8608c4-2369-4b74-927a-00eea2f1ca6e-kube-api-access-jlql6\") pod \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.329927 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-config-data\") pod \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.330087 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-combined-ca-bundle\") pod \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\" (UID: \"3a8608c4-2369-4b74-927a-00eea2f1ca6e\") " Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.340316 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8608c4-2369-4b74-927a-00eea2f1ca6e-kube-api-access-jlql6" (OuterVolumeSpecName: "kube-api-access-jlql6") pod "3a8608c4-2369-4b74-927a-00eea2f1ca6e" (UID: "3a8608c4-2369-4b74-927a-00eea2f1ca6e"). InnerVolumeSpecName "kube-api-access-jlql6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.379242 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-config-data" (OuterVolumeSpecName: "config-data") pod "3a8608c4-2369-4b74-927a-00eea2f1ca6e" (UID: "3a8608c4-2369-4b74-927a-00eea2f1ca6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.388086 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a8608c4-2369-4b74-927a-00eea2f1ca6e" (UID: "3a8608c4-2369-4b74-927a-00eea2f1ca6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.435158 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.435201 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlql6\" (UniqueName: \"kubernetes.io/projected/3a8608c4-2369-4b74-927a-00eea2f1ca6e-kube-api-access-jlql6\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.435216 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8608c4-2369-4b74-927a-00eea2f1ca6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:25 crc kubenswrapper[4710]: I1002 13:11:25.568494 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b87b8bb9-9pdwf"] Oct 02 13:11:25 crc kubenswrapper[4710]: W1002 13:11:25.574947 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93d26107_27ce_422c_b17d_ae46b170f8b2.slice/crio-0f0fb2eb9a93f69d59c2c6d16af450187e8108daa7d09f554bd0c06dee0be8df WatchSource:0}: Error finding container 0f0fb2eb9a93f69d59c2c6d16af450187e8108daa7d09f554bd0c06dee0be8df: Status 404 returned error can't find the container with id 0f0fb2eb9a93f69d59c2c6d16af450187e8108daa7d09f554bd0c06dee0be8df Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.195292 4710 generic.go:334] "Generic (PLEG): container finished" podID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerID="5cdf0f5de9a003e9d34615be5237e879c00c84c25a2037527087099e08f637de" exitCode=0 Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.195645 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" event={"ID":"93d26107-27ce-422c-b17d-ae46b170f8b2","Type":"ContainerDied","Data":"5cdf0f5de9a003e9d34615be5237e879c00c84c25a2037527087099e08f637de"} Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.195678 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" event={"ID":"93d26107-27ce-422c-b17d-ae46b170f8b2","Type":"ContainerStarted","Data":"0f0fb2eb9a93f69d59c2c6d16af450187e8108daa7d09f554bd0c06dee0be8df"} Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.203339 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.203960 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a8608c4-2369-4b74-927a-00eea2f1ca6e","Type":"ContainerDied","Data":"181a54f63b2b177c3a27f3b2cf4108572ce00d1007f1cf3c3948afbda784efba"} Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.204016 4710 scope.go:117] "RemoveContainer" containerID="0c5aba13999b99e49da83eb41df108889038dff1ec92b039f7873ea54061e1b6" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.415177 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.427307 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.440698 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:11:26 crc kubenswrapper[4710]: E1002 13:11:26.441165 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8608c4-2369-4b74-927a-00eea2f1ca6e" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.441189 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8608c4-2369-4b74-927a-00eea2f1ca6e" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.441450 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8608c4-2369-4b74-927a-00eea2f1ca6e" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.442258 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.445379 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.451411 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.452367 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.452494 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.478687 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.478777 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.478919 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wb8t\" (UniqueName: \"kubernetes.io/projected/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-kube-api-access-4wb8t\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.478943 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.478974 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.590626 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.591396 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.592636 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wb8t\" (UniqueName: \"kubernetes.io/projected/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-kube-api-access-4wb8t\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.592710 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.592851 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.597978 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.605496 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.605496 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.606025 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.618040 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wb8t\" (UniqueName: \"kubernetes.io/projected/5fd38591-5a16-4ce9-89a0-8d7b94fd0e22-kube-api-access-4wb8t\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.763478 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:26 crc kubenswrapper[4710]: I1002 13:11:26.942732 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a8608c4-2369-4b74-927a-00eea2f1ca6e" path="/var/lib/kubelet/pods/3a8608c4-2369-4b74-927a-00eea2f1ca6e/volumes" Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.105068 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.214224 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" event={"ID":"93d26107-27ce-422c-b17d-ae46b170f8b2","Type":"ContainerStarted","Data":"15e2a39a6cf9ceab890ac8e1c35cc9fe9369ab8cd4104773e6d068b4285bf489"} Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.214497 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.216434 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-log" containerID="cri-o://5cfb918a1b80d50a17cb2bf2e97f5fafcdeebf4d9ad359a4449fcd4a9d0fbf30" gracePeriod=30 Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.216448 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-api" containerID="cri-o://221bb79a48952b1ae9932896c64fe61d69901d4b080df376ffb34d1129fa647e" gracePeriod=30 Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.237435 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" podStartSLOduration=3.237411176 podStartE2EDuration="3.237411176s" podCreationTimestamp="2025-10-02 13:11:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:27.232568074 +0000 UTC m=+1331.338978957" watchObservedRunningTime="2025-10-02 13:11:27.237411176 +0000 UTC m=+1331.343822079" Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.271291 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.520290 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.521013 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-central-agent" containerID="cri-o://cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651" gracePeriod=30 Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.521097 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="proxy-httpd" containerID="cri-o://32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e" gracePeriod=30 Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.521151 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="sg-core" containerID="cri-o://c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6" gracePeriod=30 Oct 02 13:11:27 crc kubenswrapper[4710]: I1002 13:11:27.521198 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-notification-agent" containerID="cri-o://a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba" gracePeriod=30 Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.226946 4710 generic.go:334] "Generic (PLEG): container finished" podID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerID="5cfb918a1b80d50a17cb2bf2e97f5fafcdeebf4d9ad359a4449fcd4a9d0fbf30" exitCode=143 Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.227017 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4eba802-f7d9-4994-89ee-c0d4ab52ab51","Type":"ContainerDied","Data":"5cfb918a1b80d50a17cb2bf2e97f5fafcdeebf4d9ad359a4449fcd4a9d0fbf30"} Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.228954 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22","Type":"ContainerStarted","Data":"2f852af1fcd8d793f0fdbb94a53b2a08750e5898b042d35269e900a7b1043839"} Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.228995 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fd38591-5a16-4ce9-89a0-8d7b94fd0e22","Type":"ContainerStarted","Data":"bc6a37419b7b46fcb0946fab34751c629cbc9e6b5e19fa57cd3713e6252db437"} Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.231608 4710 generic.go:334] "Generic (PLEG): container finished" podID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerID="32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e" exitCode=0 Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.231637 4710 generic.go:334] "Generic (PLEG): container finished" podID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerID="c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6" exitCode=2 Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.231648 4710 generic.go:334] "Generic (PLEG): container finished" podID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerID="cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651" exitCode=0 Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.231705 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerDied","Data":"32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e"} Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.231783 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerDied","Data":"c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6"} Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.231798 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerDied","Data":"cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651"} Oct 02 13:11:28 crc kubenswrapper[4710]: I1002 13:11:28.260225 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.260197465 podStartE2EDuration="2.260197465s" podCreationTimestamp="2025-10-02 13:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:28.247843013 +0000 UTC m=+1332.354253926" watchObservedRunningTime="2025-10-02 13:11:28.260197465 +0000 UTC m=+1332.366608348" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.247362 4710 generic.go:334] "Generic (PLEG): container finished" podID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerID="221bb79a48952b1ae9932896c64fe61d69901d4b080df376ffb34d1129fa647e" exitCode=0 Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.247535 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4eba802-f7d9-4994-89ee-c0d4ab52ab51","Type":"ContainerDied","Data":"221bb79a48952b1ae9932896c64fe61d69901d4b080df376ffb34d1129fa647e"} Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.351692 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.450700 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-combined-ca-bundle\") pod \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.451198 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-logs" (OuterVolumeSpecName: "logs") pod "b4eba802-f7d9-4994-89ee-c0d4ab52ab51" (UID: "b4eba802-f7d9-4994-89ee-c0d4ab52ab51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.453734 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-logs\") pod \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.453838 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-config-data\") pod \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.453974 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mw89\" (UniqueName: \"kubernetes.io/projected/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-kube-api-access-2mw89\") pod \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\" (UID: \"b4eba802-f7d9-4994-89ee-c0d4ab52ab51\") " Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.455643 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.488117 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-kube-api-access-2mw89" (OuterVolumeSpecName: "kube-api-access-2mw89") pod "b4eba802-f7d9-4994-89ee-c0d4ab52ab51" (UID: "b4eba802-f7d9-4994-89ee-c0d4ab52ab51"). InnerVolumeSpecName "kube-api-access-2mw89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.501593 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-config-data" (OuterVolumeSpecName: "config-data") pod "b4eba802-f7d9-4994-89ee-c0d4ab52ab51" (UID: "b4eba802-f7d9-4994-89ee-c0d4ab52ab51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.526593 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4eba802-f7d9-4994-89ee-c0d4ab52ab51" (UID: "b4eba802-f7d9-4994-89ee-c0d4ab52ab51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.558515 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mw89\" (UniqueName: \"kubernetes.io/projected/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-kube-api-access-2mw89\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.558561 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:29 crc kubenswrapper[4710]: I1002 13:11:29.558575 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4eba802-f7d9-4994-89ee-c0d4ab52ab51-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.259352 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4eba802-f7d9-4994-89ee-c0d4ab52ab51","Type":"ContainerDied","Data":"328c6a527d08438877ac01f0598724bd8d298f8d1fdf28ee185ea3f290c04f55"} Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.259438 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.259727 4710 scope.go:117] "RemoveContainer" containerID="221bb79a48952b1ae9932896c64fe61d69901d4b080df376ffb34d1129fa647e" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.288815 4710 scope.go:117] "RemoveContainer" containerID="5cfb918a1b80d50a17cb2bf2e97f5fafcdeebf4d9ad359a4449fcd4a9d0fbf30" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.313808 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.320583 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.372093 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:30 crc kubenswrapper[4710]: E1002 13:11:30.372627 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-api" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.372645 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-api" Oct 02 13:11:30 crc kubenswrapper[4710]: E1002 13:11:30.372660 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-log" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.372667 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-log" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.372946 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-api" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.372992 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" containerName="nova-api-log" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.375896 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.378948 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.379148 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.379325 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.390717 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.475842 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-public-tls-certs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.476042 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.476221 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-config-data\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.476455 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381e640f-1570-4d2f-876f-f0a36147712e-logs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.476654 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm458\" (UniqueName: \"kubernetes.io/projected/381e640f-1570-4d2f-876f-f0a36147712e-kube-api-access-tm458\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.476766 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.578376 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381e640f-1570-4d2f-876f-f0a36147712e-logs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.578890 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm458\" (UniqueName: \"kubernetes.io/projected/381e640f-1570-4d2f-876f-f0a36147712e-kube-api-access-tm458\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.578926 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381e640f-1570-4d2f-876f-f0a36147712e-logs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.578939 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.579116 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-public-tls-certs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.579222 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.579389 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-config-data\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.585578 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-public-tls-certs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.585809 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-config-data\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.585996 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.598041 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.601451 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm458\" (UniqueName: \"kubernetes.io/projected/381e640f-1570-4d2f-876f-f0a36147712e-kube-api-access-tm458\") pod \"nova-api-0\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.708391 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.769295 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.886958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-scripts\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.887713 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-config-data\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888110 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-ceilometer-tls-certs\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888185 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-log-httpd\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888215 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-sg-core-conf-yaml\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888313 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6tgm\" (UniqueName: \"kubernetes.io/projected/ab18b551-f6ad-41b5-bd92-b98701675ff6-kube-api-access-x6tgm\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888355 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-combined-ca-bundle\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888509 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-run-httpd\") pod \"ab18b551-f6ad-41b5-bd92-b98701675ff6\" (UID: \"ab18b551-f6ad-41b5-bd92-b98701675ff6\") " Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.888910 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.889675 4710 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.890504 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.895902 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-scripts" (OuterVolumeSpecName: "scripts") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.897313 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab18b551-f6ad-41b5-bd92-b98701675ff6-kube-api-access-x6tgm" (OuterVolumeSpecName: "kube-api-access-x6tgm") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "kube-api-access-x6tgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.922022 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4eba802-f7d9-4994-89ee-c0d4ab52ab51" path="/var/lib/kubelet/pods/b4eba802-f7d9-4994-89ee-c0d4ab52ab51/volumes" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.932612 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.955045 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.991778 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6tgm\" (UniqueName: \"kubernetes.io/projected/ab18b551-f6ad-41b5-bd92-b98701675ff6-kube-api-access-x6tgm\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.991815 4710 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab18b551-f6ad-41b5-bd92-b98701675ff6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.991827 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.991839 4710 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.991851 4710 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:30 crc kubenswrapper[4710]: I1002 13:11:30.995174 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.048696 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-config-data" (OuterVolumeSpecName: "config-data") pod "ab18b551-f6ad-41b5-bd92-b98701675ff6" (UID: "ab18b551-f6ad-41b5-bd92-b98701675ff6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.093823 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.093866 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab18b551-f6ad-41b5-bd92-b98701675ff6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.199039 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:31 crc kubenswrapper[4710]: W1002 13:11:31.201395 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod381e640f_1570_4d2f_876f_f0a36147712e.slice/crio-3fa35b75120445542de9d554a2997da47a52f61daf4a4cb939a92887fac3578a WatchSource:0}: Error finding container 3fa35b75120445542de9d554a2997da47a52f61daf4a4cb939a92887fac3578a: Status 404 returned error can't find the container with id 3fa35b75120445542de9d554a2997da47a52f61daf4a4cb939a92887fac3578a Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.272104 4710 generic.go:334] "Generic (PLEG): container finished" podID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerID="a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba" exitCode=0 Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.272173 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.272243 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerDied","Data":"a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba"} Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.272333 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab18b551-f6ad-41b5-bd92-b98701675ff6","Type":"ContainerDied","Data":"43cd86de84ed02da06acc596f78e0f11d5df141b968360c71ede366b352c09d0"} Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.272360 4710 scope.go:117] "RemoveContainer" containerID="32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.273878 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381e640f-1570-4d2f-876f-f0a36147712e","Type":"ContainerStarted","Data":"3fa35b75120445542de9d554a2997da47a52f61daf4a4cb939a92887fac3578a"} Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.338171 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.356142 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.359409 4710 scope.go:117] "RemoveContainer" containerID="c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.371539 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.372015 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-central-agent" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372027 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-central-agent" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.372043 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="sg-core" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372049 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="sg-core" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.372065 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="proxy-httpd" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372071 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="proxy-httpd" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.372079 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-notification-agent" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372085 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-notification-agent" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372266 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-notification-agent" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372287 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="sg-core" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372310 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="ceilometer-central-agent" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.372317 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" containerName="proxy-httpd" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.374083 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.377260 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.377260 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.385190 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.398676 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bd2ecce-524b-4abd-adfb-effdd3cc8204-run-httpd\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.398738 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzqqr\" (UniqueName: \"kubernetes.io/projected/9bd2ecce-524b-4abd-adfb-effdd3cc8204-kube-api-access-nzqqr\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.398790 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.398831 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-scripts\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.398855 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.399006 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.399035 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bd2ecce-524b-4abd-adfb-effdd3cc8204-log-httpd\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.399058 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-config-data\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.402301 4710 scope.go:117] "RemoveContainer" containerID="a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.407687 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.431055 4710 scope.go:117] "RemoveContainer" containerID="cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.461905 4710 scope.go:117] "RemoveContainer" containerID="32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.462327 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e\": container with ID starting with 32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e not found: ID does not exist" containerID="32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.462362 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e"} err="failed to get container status \"32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e\": rpc error: code = NotFound desc = could not find container \"32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e\": container with ID starting with 32cbc9f2a382e1af81f282f2b0fd59992f2b13c094e72e9ddd68bb9b99bbd87e not found: ID does not exist" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.462385 4710 scope.go:117] "RemoveContainer" containerID="c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.462720 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6\": container with ID starting with c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6 not found: ID does not exist" containerID="c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.462827 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6"} err="failed to get container status \"c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6\": rpc error: code = NotFound desc = could not find container \"c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6\": container with ID starting with c4ad4dba723c01af6bd23335f10e03d749a344a447345c60d8586a40b86c00a6 not found: ID does not exist" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.462849 4710 scope.go:117] "RemoveContainer" containerID="a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.463118 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba\": container with ID starting with a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba not found: ID does not exist" containerID="a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.463144 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba"} err="failed to get container status \"a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba\": rpc error: code = NotFound desc = could not find container \"a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba\": container with ID starting with a7cae8734823097c42b417f833f55343112f1c0bba949f8ab472cc5928f856ba not found: ID does not exist" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.463162 4710 scope.go:117] "RemoveContainer" containerID="cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651" Oct 02 13:11:31 crc kubenswrapper[4710]: E1002 13:11:31.463481 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651\": container with ID starting with cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651 not found: ID does not exist" containerID="cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.463506 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651"} err="failed to get container status \"cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651\": rpc error: code = NotFound desc = could not find container \"cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651\": container with ID starting with cc518c1573689e24cccc79777def461740c863c5d37597067b8854710efcf651 not found: ID does not exist" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.499975 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bd2ecce-524b-4abd-adfb-effdd3cc8204-run-httpd\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500026 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzqqr\" (UniqueName: \"kubernetes.io/projected/9bd2ecce-524b-4abd-adfb-effdd3cc8204-kube-api-access-nzqqr\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500071 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500094 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-scripts\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500107 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500284 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500348 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bd2ecce-524b-4abd-adfb-effdd3cc8204-log-httpd\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500367 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-config-data\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500597 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bd2ecce-524b-4abd-adfb-effdd3cc8204-run-httpd\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.500889 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bd2ecce-524b-4abd-adfb-effdd3cc8204-log-httpd\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.503690 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.504633 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.505237 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.505476 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-config-data\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.505629 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bd2ecce-524b-4abd-adfb-effdd3cc8204-scripts\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.517874 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzqqr\" (UniqueName: \"kubernetes.io/projected/9bd2ecce-524b-4abd-adfb-effdd3cc8204-kube-api-access-nzqqr\") pod \"ceilometer-0\" (UID: \"9bd2ecce-524b-4abd-adfb-effdd3cc8204\") " pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.702562 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 13:11:31 crc kubenswrapper[4710]: I1002 13:11:31.764054 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:32 crc kubenswrapper[4710]: I1002 13:11:32.180086 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 13:11:32 crc kubenswrapper[4710]: I1002 13:11:32.289226 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381e640f-1570-4d2f-876f-f0a36147712e","Type":"ContainerStarted","Data":"8ae2c5af452eb6f99b594bcb79138f60c23efeabaff974bc313665b7d70a4e78"} Oct 02 13:11:32 crc kubenswrapper[4710]: I1002 13:11:32.289633 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381e640f-1570-4d2f-876f-f0a36147712e","Type":"ContainerStarted","Data":"98d2bc144381e5beb281c0d266cbaa5efaa7b0353d8d56c0b78bae65ade86805"} Oct 02 13:11:32 crc kubenswrapper[4710]: I1002 13:11:32.290392 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bd2ecce-524b-4abd-adfb-effdd3cc8204","Type":"ContainerStarted","Data":"e7c55f34a7a76ca8ef9751c6da43dfde0f0f59b9a6af5ae59179699904c443b6"} Oct 02 13:11:32 crc kubenswrapper[4710]: I1002 13:11:32.330675 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.330652598 podStartE2EDuration="2.330652598s" podCreationTimestamp="2025-10-02 13:11:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:32.328073003 +0000 UTC m=+1336.434483886" watchObservedRunningTime="2025-10-02 13:11:32.330652598 +0000 UTC m=+1336.437063481" Oct 02 13:11:32 crc kubenswrapper[4710]: I1002 13:11:32.925194 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab18b551-f6ad-41b5-bd92-b98701675ff6" path="/var/lib/kubelet/pods/ab18b551-f6ad-41b5-bd92-b98701675ff6/volumes" Oct 02 13:11:33 crc kubenswrapper[4710]: I1002 13:11:33.300981 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bd2ecce-524b-4abd-adfb-effdd3cc8204","Type":"ContainerStarted","Data":"9ac7e49f7a586739d193c628af96dd42a4d8b524f8ad5c1fbfb236ec1c1b8356"} Oct 02 13:11:33 crc kubenswrapper[4710]: I1002 13:11:33.301024 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bd2ecce-524b-4abd-adfb-effdd3cc8204","Type":"ContainerStarted","Data":"84fd2538b8c6328c491158f53d81c12472f5b5fde79693d011c51318087be11e"} Oct 02 13:11:34 crc kubenswrapper[4710]: I1002 13:11:34.315693 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bd2ecce-524b-4abd-adfb-effdd3cc8204","Type":"ContainerStarted","Data":"9fd718410d702443baab96468f49f5b9625a1def532e9346ec7e610e842fccd5"} Oct 02 13:11:34 crc kubenswrapper[4710]: I1002 13:11:34.839983 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:11:34 crc kubenswrapper[4710]: I1002 13:11:34.934293 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b455c8df9-g6bzr"] Oct 02 13:11:34 crc kubenswrapper[4710]: I1002 13:11:34.934586 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerName="dnsmasq-dns" containerID="cri-o://64db8b97a25ab0c2628ad9533e43803a084b5f3ae23dabfc0942ae00ce911995" gracePeriod=10 Oct 02 13:11:34 crc kubenswrapper[4710]: I1002 13:11:34.956142 4710 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod21d98a36-dd9d-42b5-b26c-a6fa30189c44"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod21d98a36-dd9d-42b5-b26c-a6fa30189c44] : Timed out while waiting for systemd to remove kubepods-besteffort-pod21d98a36_dd9d_42b5_b26c_a6fa30189c44.slice" Oct 02 13:11:34 crc kubenswrapper[4710]: E1002 13:11:34.956206 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod21d98a36-dd9d-42b5-b26c-a6fa30189c44] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod21d98a36-dd9d-42b5-b26c-a6fa30189c44] : Timed out while waiting for systemd to remove kubepods-besteffort-pod21d98a36_dd9d_42b5_b26c_a6fa30189c44.slice" pod="openstack/nova-cell1-conductor-db-sync-djnzg" podUID="21d98a36-dd9d-42b5-b26c-a6fa30189c44" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.342531 4710 generic.go:334] "Generic (PLEG): container finished" podID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerID="64db8b97a25ab0c2628ad9533e43803a084b5f3ae23dabfc0942ae00ce911995" exitCode=0 Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.343162 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djnzg" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.343533 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" event={"ID":"237922f1-0c3a-4d6f-9404-2a9fe3efa393","Type":"ContainerDied","Data":"64db8b97a25ab0c2628ad9533e43803a084b5f3ae23dabfc0942ae00ce911995"} Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.495635 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.587660 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-sb\") pod \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.587759 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-svc\") pod \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.587800 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lptfl\" (UniqueName: \"kubernetes.io/projected/237922f1-0c3a-4d6f-9404-2a9fe3efa393-kube-api-access-lptfl\") pod \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.587842 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-swift-storage-0\") pod \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.587980 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-nb\") pod \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.588090 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-config\") pod \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\" (UID: \"237922f1-0c3a-4d6f-9404-2a9fe3efa393\") " Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.619143 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237922f1-0c3a-4d6f-9404-2a9fe3efa393-kube-api-access-lptfl" (OuterVolumeSpecName: "kube-api-access-lptfl") pod "237922f1-0c3a-4d6f-9404-2a9fe3efa393" (UID: "237922f1-0c3a-4d6f-9404-2a9fe3efa393"). InnerVolumeSpecName "kube-api-access-lptfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.655446 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "237922f1-0c3a-4d6f-9404-2a9fe3efa393" (UID: "237922f1-0c3a-4d6f-9404-2a9fe3efa393"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.655502 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "237922f1-0c3a-4d6f-9404-2a9fe3efa393" (UID: "237922f1-0c3a-4d6f-9404-2a9fe3efa393"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.657958 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "237922f1-0c3a-4d6f-9404-2a9fe3efa393" (UID: "237922f1-0c3a-4d6f-9404-2a9fe3efa393"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.661329 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-config" (OuterVolumeSpecName: "config") pod "237922f1-0c3a-4d6f-9404-2a9fe3efa393" (UID: "237922f1-0c3a-4d6f-9404-2a9fe3efa393"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.673812 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "237922f1-0c3a-4d6f-9404-2a9fe3efa393" (UID: "237922f1-0c3a-4d6f-9404-2a9fe3efa393"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.690621 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.690659 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.690671 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lptfl\" (UniqueName: \"kubernetes.io/projected/237922f1-0c3a-4d6f-9404-2a9fe3efa393-kube-api-access-lptfl\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.690683 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.690691 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:35 crc kubenswrapper[4710]: I1002 13:11:35.690701 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237922f1-0c3a-4d6f-9404-2a9fe3efa393-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.354139 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" event={"ID":"237922f1-0c3a-4d6f-9404-2a9fe3efa393","Type":"ContainerDied","Data":"1b612905e9a1f55fc10aabee6482a868e4f03797b0fbc0a9bbf0cbe377aae9dd"} Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.354167 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b455c8df9-g6bzr" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.355613 4710 scope.go:117] "RemoveContainer" containerID="64db8b97a25ab0c2628ad9533e43803a084b5f3ae23dabfc0942ae00ce911995" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.356552 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bd2ecce-524b-4abd-adfb-effdd3cc8204","Type":"ContainerStarted","Data":"d53d01b55dd636a1d6ee8329f748be20b8b8f5f2685d36c0e749fd4b80479d02"} Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.356703 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.376385 4710 scope.go:117] "RemoveContainer" containerID="332a5602d52bf3b1a2204ab29137ee19e9f8f589a06a4a6540e7af05212ab2d8" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.404522 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.753292512 podStartE2EDuration="5.404505637s" podCreationTimestamp="2025-10-02 13:11:31 +0000 UTC" firstStartedPulling="2025-10-02 13:11:32.194866809 +0000 UTC m=+1336.301277692" lastFinishedPulling="2025-10-02 13:11:35.846079934 +0000 UTC m=+1339.952490817" observedRunningTime="2025-10-02 13:11:36.400582957 +0000 UTC m=+1340.506993840" watchObservedRunningTime="2025-10-02 13:11:36.404505637 +0000 UTC m=+1340.510916520" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.421445 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b455c8df9-g6bzr"] Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.429453 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b455c8df9-g6bzr"] Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.763722 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.786723 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:36 crc kubenswrapper[4710]: I1002 13:11:36.915491 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" path="/var/lib/kubelet/pods/237922f1-0c3a-4d6f-9404-2a9fe3efa393/volumes" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.400004 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.593152 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-grpct"] Oct 02 13:11:37 crc kubenswrapper[4710]: E1002 13:11:37.593548 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerName="dnsmasq-dns" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.593564 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerName="dnsmasq-dns" Oct 02 13:11:37 crc kubenswrapper[4710]: E1002 13:11:37.593602 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerName="init" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.593816 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerName="init" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.595336 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="237922f1-0c3a-4d6f-9404-2a9fe3efa393" containerName="dnsmasq-dns" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.596059 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.598512 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.598679 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.620067 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-grpct"] Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.731706 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wxbm\" (UniqueName: \"kubernetes.io/projected/9cc732de-8466-4872-a4ce-7fc91a64dd8d-kube-api-access-7wxbm\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.732025 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.732164 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-scripts\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.732262 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-config-data\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.833625 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wxbm\" (UniqueName: \"kubernetes.io/projected/9cc732de-8466-4872-a4ce-7fc91a64dd8d-kube-api-access-7wxbm\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.833704 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.833843 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-scripts\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.833907 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-config-data\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.840531 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-config-data\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.841560 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-scripts\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.846921 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.856329 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wxbm\" (UniqueName: \"kubernetes.io/projected/9cc732de-8466-4872-a4ce-7fc91a64dd8d-kube-api-access-7wxbm\") pod \"nova-cell1-cell-mapping-grpct\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:37 crc kubenswrapper[4710]: I1002 13:11:37.917414 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:38 crc kubenswrapper[4710]: I1002 13:11:38.401964 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-grpct"] Oct 02 13:11:39 crc kubenswrapper[4710]: I1002 13:11:39.393326 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-grpct" event={"ID":"9cc732de-8466-4872-a4ce-7fc91a64dd8d","Type":"ContainerStarted","Data":"04a80ba2e91c8740f43c25de43298edef39174bff81dbe4ebde05e6beeb84515"} Oct 02 13:11:39 crc kubenswrapper[4710]: I1002 13:11:39.393740 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-grpct" event={"ID":"9cc732de-8466-4872-a4ce-7fc91a64dd8d","Type":"ContainerStarted","Data":"ae7aef4c7d824a5c985bf481e5626328a5b371579f27f5856bb84c6f06f0136d"} Oct 02 13:11:39 crc kubenswrapper[4710]: I1002 13:11:39.412370 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-grpct" podStartSLOduration=2.412346735 podStartE2EDuration="2.412346735s" podCreationTimestamp="2025-10-02 13:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:39.410901448 +0000 UTC m=+1343.517312331" watchObservedRunningTime="2025-10-02 13:11:39.412346735 +0000 UTC m=+1343.518757618" Oct 02 13:11:40 crc kubenswrapper[4710]: I1002 13:11:40.709665 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:11:40 crc kubenswrapper[4710]: I1002 13:11:40.710467 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:11:41 crc kubenswrapper[4710]: I1002 13:11:41.721919 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:41 crc kubenswrapper[4710]: I1002 13:11:41.721937 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:44 crc kubenswrapper[4710]: I1002 13:11:44.446681 4710 generic.go:334] "Generic (PLEG): container finished" podID="9cc732de-8466-4872-a4ce-7fc91a64dd8d" containerID="04a80ba2e91c8740f43c25de43298edef39174bff81dbe4ebde05e6beeb84515" exitCode=0 Oct 02 13:11:44 crc kubenswrapper[4710]: I1002 13:11:44.446739 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-grpct" event={"ID":"9cc732de-8466-4872-a4ce-7fc91a64dd8d","Type":"ContainerDied","Data":"04a80ba2e91c8740f43c25de43298edef39174bff81dbe4ebde05e6beeb84515"} Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.853921 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.898962 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-combined-ca-bundle\") pod \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.899043 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wxbm\" (UniqueName: \"kubernetes.io/projected/9cc732de-8466-4872-a4ce-7fc91a64dd8d-kube-api-access-7wxbm\") pod \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.899208 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-config-data\") pod \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.899240 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-scripts\") pod \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\" (UID: \"9cc732de-8466-4872-a4ce-7fc91a64dd8d\") " Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.907623 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cc732de-8466-4872-a4ce-7fc91a64dd8d-kube-api-access-7wxbm" (OuterVolumeSpecName: "kube-api-access-7wxbm") pod "9cc732de-8466-4872-a4ce-7fc91a64dd8d" (UID: "9cc732de-8466-4872-a4ce-7fc91a64dd8d"). InnerVolumeSpecName "kube-api-access-7wxbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.907733 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-scripts" (OuterVolumeSpecName: "scripts") pod "9cc732de-8466-4872-a4ce-7fc91a64dd8d" (UID: "9cc732de-8466-4872-a4ce-7fc91a64dd8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.931075 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cc732de-8466-4872-a4ce-7fc91a64dd8d" (UID: "9cc732de-8466-4872-a4ce-7fc91a64dd8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:45 crc kubenswrapper[4710]: I1002 13:11:45.931985 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-config-data" (OuterVolumeSpecName: "config-data") pod "9cc732de-8466-4872-a4ce-7fc91a64dd8d" (UID: "9cc732de-8466-4872-a4ce-7fc91a64dd8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.001564 4710 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.001598 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.001608 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wxbm\" (UniqueName: \"kubernetes.io/projected/9cc732de-8466-4872-a4ce-7fc91a64dd8d-kube-api-access-7wxbm\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.001628 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc732de-8466-4872-a4ce-7fc91a64dd8d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.470026 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-grpct" event={"ID":"9cc732de-8466-4872-a4ce-7fc91a64dd8d","Type":"ContainerDied","Data":"ae7aef4c7d824a5c985bf481e5626328a5b371579f27f5856bb84c6f06f0136d"} Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.470353 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae7aef4c7d824a5c985bf481e5626328a5b371579f27f5856bb84c6f06f0136d" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.470128 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-grpct" Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.658778 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.659272 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-log" containerID="cri-o://98d2bc144381e5beb281c0d266cbaa5efaa7b0353d8d56c0b78bae65ade86805" gracePeriod=30 Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.659497 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-api" containerID="cri-o://8ae2c5af452eb6f99b594bcb79138f60c23efeabaff974bc313665b7d70a4e78" gracePeriod=30 Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.673320 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.673531 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" containerName="nova-scheduler-scheduler" containerID="cri-o://4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9" gracePeriod=30 Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.687128 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.687367 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-log" containerID="cri-o://e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c" gracePeriod=30 Oct 02 13:11:46 crc kubenswrapper[4710]: I1002 13:11:46.687911 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-metadata" containerID="cri-o://400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a" gracePeriod=30 Oct 02 13:11:47 crc kubenswrapper[4710]: I1002 13:11:47.482666 4710 generic.go:334] "Generic (PLEG): container finished" podID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerID="e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c" exitCode=143 Oct 02 13:11:47 crc kubenswrapper[4710]: I1002 13:11:47.482759 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44fd49c6-fa11-4c3a-9a95-72ae18059ac7","Type":"ContainerDied","Data":"e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c"} Oct 02 13:11:47 crc kubenswrapper[4710]: I1002 13:11:47.485082 4710 generic.go:334] "Generic (PLEG): container finished" podID="381e640f-1570-4d2f-876f-f0a36147712e" containerID="98d2bc144381e5beb281c0d266cbaa5efaa7b0353d8d56c0b78bae65ade86805" exitCode=143 Oct 02 13:11:47 crc kubenswrapper[4710]: I1002 13:11:47.485153 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381e640f-1570-4d2f-876f-f0a36147712e","Type":"ContainerDied","Data":"98d2bc144381e5beb281c0d266cbaa5efaa7b0353d8d56c0b78bae65ade86805"} Oct 02 13:11:47 crc kubenswrapper[4710]: I1002 13:11:47.556362 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": read tcp 10.217.0.2:45694->10.217.0.214:8775: read: connection reset by peer" Oct 02 13:11:47 crc kubenswrapper[4710]: I1002 13:11:47.556419 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": read tcp 10.217.0.2:45692->10.217.0.214:8775: read: connection reset by peer" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.068525 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.150184 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-logs\") pod \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.150259 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-config-data\") pod \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.150299 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-nova-metadata-tls-certs\") pod \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.150360 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgnmw\" (UniqueName: \"kubernetes.io/projected/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-kube-api-access-kgnmw\") pod \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.150425 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-combined-ca-bundle\") pod \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\" (UID: \"44fd49c6-fa11-4c3a-9a95-72ae18059ac7\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.150599 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-logs" (OuterVolumeSpecName: "logs") pod "44fd49c6-fa11-4c3a-9a95-72ae18059ac7" (UID: "44fd49c6-fa11-4c3a-9a95-72ae18059ac7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.153982 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.157227 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-kube-api-access-kgnmw" (OuterVolumeSpecName: "kube-api-access-kgnmw") pod "44fd49c6-fa11-4c3a-9a95-72ae18059ac7" (UID: "44fd49c6-fa11-4c3a-9a95-72ae18059ac7"). InnerVolumeSpecName "kube-api-access-kgnmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.187373 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-config-data" (OuterVolumeSpecName: "config-data") pod "44fd49c6-fa11-4c3a-9a95-72ae18059ac7" (UID: "44fd49c6-fa11-4c3a-9a95-72ae18059ac7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.194119 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44fd49c6-fa11-4c3a-9a95-72ae18059ac7" (UID: "44fd49c6-fa11-4c3a-9a95-72ae18059ac7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.215357 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "44fd49c6-fa11-4c3a-9a95-72ae18059ac7" (UID: "44fd49c6-fa11-4c3a-9a95-72ae18059ac7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.255862 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.255898 4710 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.255910 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgnmw\" (UniqueName: \"kubernetes.io/projected/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-kube-api-access-kgnmw\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.255918 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44fd49c6-fa11-4c3a-9a95-72ae18059ac7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.519786 4710 generic.go:334] "Generic (PLEG): container finished" podID="381e640f-1570-4d2f-876f-f0a36147712e" containerID="8ae2c5af452eb6f99b594bcb79138f60c23efeabaff974bc313665b7d70a4e78" exitCode=0 Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.520003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381e640f-1570-4d2f-876f-f0a36147712e","Type":"ContainerDied","Data":"8ae2c5af452eb6f99b594bcb79138f60c23efeabaff974bc313665b7d70a4e78"} Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.523831 4710 generic.go:334] "Generic (PLEG): container finished" podID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerID="400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a" exitCode=0 Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.523873 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.523872 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44fd49c6-fa11-4c3a-9a95-72ae18059ac7","Type":"ContainerDied","Data":"400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a"} Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.524021 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"44fd49c6-fa11-4c3a-9a95-72ae18059ac7","Type":"ContainerDied","Data":"8bb9774bfe20fe64651fa783617b3fc8c7e691be108f10175ee75669999da20c"} Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.524072 4710 scope.go:117] "RemoveContainer" containerID="400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.566046 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.605486 4710 scope.go:117] "RemoveContainer" containerID="e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.607793 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.619401 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:48 crc kubenswrapper[4710]: E1002 13:11:48.620275 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc732de-8466-4872-a4ce-7fc91a64dd8d" containerName="nova-manage" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.620357 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc732de-8466-4872-a4ce-7fc91a64dd8d" containerName="nova-manage" Oct 02 13:11:48 crc kubenswrapper[4710]: E1002 13:11:48.620433 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-log" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.620492 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-log" Oct 02 13:11:48 crc kubenswrapper[4710]: E1002 13:11:48.620561 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-metadata" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.620622 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-metadata" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.621053 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-log" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.621215 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cc732de-8466-4872-a4ce-7fc91a64dd8d" containerName="nova-manage" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.621294 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" containerName="nova-metadata-metadata" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.623170 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.626232 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.626507 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.639362 4710 scope.go:117] "RemoveContainer" containerID="400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a" Oct 02 13:11:48 crc kubenswrapper[4710]: E1002 13:11:48.643050 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a\": container with ID starting with 400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a not found: ID does not exist" containerID="400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.643104 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a"} err="failed to get container status \"400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a\": rpc error: code = NotFound desc = could not find container \"400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a\": container with ID starting with 400bd59e140d22d234116b5f7a14a165d46b90a63ef9859399cfbce8db82812a not found: ID does not exist" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.643134 4710 scope.go:117] "RemoveContainer" containerID="e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c" Oct 02 13:11:48 crc kubenswrapper[4710]: E1002 13:11:48.652385 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c\": container with ID starting with e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c not found: ID does not exist" containerID="e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.652464 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c"} err="failed to get container status \"e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c\": rpc error: code = NotFound desc = could not find container \"e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c\": container with ID starting with e3188f7ae1409737059ef8e4a46c5d845bf8f9d6d9ab4c011182e41b6c5ab96c not found: ID does not exist" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.660870 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.663615 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-config-data\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.663762 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smzm4\" (UniqueName: \"kubernetes.io/projected/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-kube-api-access-smzm4\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.663848 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.663873 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-logs\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.663913 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.765569 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.765685 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-config-data\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.765772 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smzm4\" (UniqueName: \"kubernetes.io/projected/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-kube-api-access-smzm4\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.765810 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.765828 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-logs\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.766332 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-logs\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.771670 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.774056 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-config-data\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.774277 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.787281 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smzm4\" (UniqueName: \"kubernetes.io/projected/e55e8319-4c74-4b16-b045-a6f37f4a1ab5-kube-api-access-smzm4\") pod \"nova-metadata-0\" (UID: \"e55e8319-4c74-4b16-b045-a6f37f4a1ab5\") " pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.795557 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.866968 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-config-data\") pod \"381e640f-1570-4d2f-876f-f0a36147712e\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.867013 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-public-tls-certs\") pod \"381e640f-1570-4d2f-876f-f0a36147712e\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.867112 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-combined-ca-bundle\") pod \"381e640f-1570-4d2f-876f-f0a36147712e\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.867145 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381e640f-1570-4d2f-876f-f0a36147712e-logs\") pod \"381e640f-1570-4d2f-876f-f0a36147712e\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.867260 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm458\" (UniqueName: \"kubernetes.io/projected/381e640f-1570-4d2f-876f-f0a36147712e-kube-api-access-tm458\") pod \"381e640f-1570-4d2f-876f-f0a36147712e\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.867312 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-internal-tls-certs\") pod \"381e640f-1570-4d2f-876f-f0a36147712e\" (UID: \"381e640f-1570-4d2f-876f-f0a36147712e\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.870664 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/381e640f-1570-4d2f-876f-f0a36147712e-logs" (OuterVolumeSpecName: "logs") pod "381e640f-1570-4d2f-876f-f0a36147712e" (UID: "381e640f-1570-4d2f-876f-f0a36147712e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.872913 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/381e640f-1570-4d2f-876f-f0a36147712e-kube-api-access-tm458" (OuterVolumeSpecName: "kube-api-access-tm458") pod "381e640f-1570-4d2f-876f-f0a36147712e" (UID: "381e640f-1570-4d2f-876f-f0a36147712e"). InnerVolumeSpecName "kube-api-access-tm458". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.886401 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.923012 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-config-data" (OuterVolumeSpecName: "config-data") pod "381e640f-1570-4d2f-876f-f0a36147712e" (UID: "381e640f-1570-4d2f-876f-f0a36147712e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.928225 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44fd49c6-fa11-4c3a-9a95-72ae18059ac7" path="/var/lib/kubelet/pods/44fd49c6-fa11-4c3a-9a95-72ae18059ac7/volumes" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.940885 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "381e640f-1570-4d2f-876f-f0a36147712e" (UID: "381e640f-1570-4d2f-876f-f0a36147712e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.961626 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.971232 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-combined-ca-bundle\") pod \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.971360 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86952\" (UniqueName: \"kubernetes.io/projected/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-kube-api-access-86952\") pod \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.971417 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-config-data\") pod \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\" (UID: \"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3\") " Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.971528 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "381e640f-1570-4d2f-876f-f0a36147712e" (UID: "381e640f-1570-4d2f-876f-f0a36147712e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.973015 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.973045 4710 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381e640f-1570-4d2f-876f-f0a36147712e-logs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.973062 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm458\" (UniqueName: \"kubernetes.io/projected/381e640f-1570-4d2f-876f-f0a36147712e-kube-api-access-tm458\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.973078 4710 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.973091 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.981208 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-kube-api-access-86952" (OuterVolumeSpecName: "kube-api-access-86952") pod "28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" (UID: "28bb35ad-6443-4ff0-86f0-358b7fe8ecc3"). InnerVolumeSpecName "kube-api-access-86952". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:11:48 crc kubenswrapper[4710]: I1002 13:11:48.988950 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "381e640f-1570-4d2f-876f-f0a36147712e" (UID: "381e640f-1570-4d2f-876f-f0a36147712e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.002894 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" (UID: "28bb35ad-6443-4ff0-86f0-358b7fe8ecc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.003302 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-config-data" (OuterVolumeSpecName: "config-data") pod "28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" (UID: "28bb35ad-6443-4ff0-86f0-358b7fe8ecc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.075311 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.075880 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86952\" (UniqueName: \"kubernetes.io/projected/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-kube-api-access-86952\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.075983 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.076060 4710 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/381e640f-1570-4d2f-876f-f0a36147712e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.493047 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.536849 4710 generic.go:334] "Generic (PLEG): container finished" podID="28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" containerID="4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9" exitCode=0 Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.536926 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.536980 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3","Type":"ContainerDied","Data":"4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9"} Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.537065 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"28bb35ad-6443-4ff0-86f0-358b7fe8ecc3","Type":"ContainerDied","Data":"a671b8be9c39cc294daa5efff68bdf38a8f7d4d158940596a7d7e28c96c21990"} Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.537098 4710 scope.go:117] "RemoveContainer" containerID="4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.542762 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381e640f-1570-4d2f-876f-f0a36147712e","Type":"ContainerDied","Data":"3fa35b75120445542de9d554a2997da47a52f61daf4a4cb939a92887fac3578a"} Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.542858 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.544505 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e55e8319-4c74-4b16-b045-a6f37f4a1ab5","Type":"ContainerStarted","Data":"c243b386440dbe9b74eed775255ec5c49356ef9caa7fd111250db5cc6cbbc33e"} Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.574775 4710 scope.go:117] "RemoveContainer" containerID="4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9" Oct 02 13:11:49 crc kubenswrapper[4710]: E1002 13:11:49.575308 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9\": container with ID starting with 4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9 not found: ID does not exist" containerID="4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.575347 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9"} err="failed to get container status \"4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9\": rpc error: code = NotFound desc = could not find container \"4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9\": container with ID starting with 4c3efe07286c801b73e89177d576962604676d709f88fcad4389f6654162dff9 not found: ID does not exist" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.575371 4710 scope.go:117] "RemoveContainer" containerID="8ae2c5af452eb6f99b594bcb79138f60c23efeabaff974bc313665b7d70a4e78" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.618281 4710 scope.go:117] "RemoveContainer" containerID="98d2bc144381e5beb281c0d266cbaa5efaa7b0353d8d56c0b78bae65ade86805" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.682453 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.704920 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.716804 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.728526 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.737311 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: E1002 13:11:49.737836 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-api" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.737858 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-api" Oct 02 13:11:49 crc kubenswrapper[4710]: E1002 13:11:49.737893 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" containerName="nova-scheduler-scheduler" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.737903 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" containerName="nova-scheduler-scheduler" Oct 02 13:11:49 crc kubenswrapper[4710]: E1002 13:11:49.737928 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-log" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.737937 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-log" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.738161 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-api" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.738180 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" containerName="nova-scheduler-scheduler" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.738211 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="381e640f-1570-4d2f-876f-f0a36147712e" containerName="nova-api-log" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.739084 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.741577 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.744345 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.752476 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.755373 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.760170 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.760453 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.761229 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.761573 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791014 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791056 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-config-data\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791076 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb9dm\" (UniqueName: \"kubernetes.io/projected/8bd1f002-8396-4368-8d3c-432a2590da5c-kube-api-access-mb9dm\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791101 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791348 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-config-data\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791599 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791793 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbnkm\" (UniqueName: \"kubernetes.io/projected/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-kube-api-access-bbnkm\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.791853 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.792202 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f002-8396-4368-8d3c-432a2590da5c-logs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.893795 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbnkm\" (UniqueName: \"kubernetes.io/projected/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-kube-api-access-bbnkm\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.893850 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.893909 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f002-8396-4368-8d3c-432a2590da5c-logs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.893980 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.894005 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-config-data\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.894027 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb9dm\" (UniqueName: \"kubernetes.io/projected/8bd1f002-8396-4368-8d3c-432a2590da5c-kube-api-access-mb9dm\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.894054 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.894098 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-config-data\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.894169 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.895002 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bd1f002-8396-4368-8d3c-432a2590da5c-logs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.898029 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.898339 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.899040 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-config-data\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.900830 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd1f002-8396-4368-8d3c-432a2590da5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.901349 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.905793 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-config-data\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.913398 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbnkm\" (UniqueName: \"kubernetes.io/projected/34e05f6e-7f38-4c5f-88fc-31be5ed5a01a-kube-api-access-bbnkm\") pod \"nova-scheduler-0\" (UID: \"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a\") " pod="openstack/nova-scheduler-0" Oct 02 13:11:49 crc kubenswrapper[4710]: I1002 13:11:49.913801 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb9dm\" (UniqueName: \"kubernetes.io/projected/8bd1f002-8396-4368-8d3c-432a2590da5c-kube-api-access-mb9dm\") pod \"nova-api-0\" (UID: \"8bd1f002-8396-4368-8d3c-432a2590da5c\") " pod="openstack/nova-api-0" Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.063830 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.087248 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 13:11:50 crc kubenswrapper[4710]: W1002 13:11:50.530054 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e05f6e_7f38_4c5f_88fc_31be5ed5a01a.slice/crio-c0819f08d134266f7d2e62a7b2010de8ab5f860ba0a3241a74a4c1e161fdfea7 WatchSource:0}: Error finding container c0819f08d134266f7d2e62a7b2010de8ab5f860ba0a3241a74a4c1e161fdfea7: Status 404 returned error can't find the container with id c0819f08d134266f7d2e62a7b2010de8ab5f860ba0a3241a74a4c1e161fdfea7 Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.533953 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.554084 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a","Type":"ContainerStarted","Data":"c0819f08d134266f7d2e62a7b2010de8ab5f860ba0a3241a74a4c1e161fdfea7"} Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.557126 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e55e8319-4c74-4b16-b045-a6f37f4a1ab5","Type":"ContainerStarted","Data":"1eebead895a3071c5db7f5c9abad0eb70d5c023447b3f2cee5fa86424ef4b51d"} Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.557160 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e55e8319-4c74-4b16-b045-a6f37f4a1ab5","Type":"ContainerStarted","Data":"b34a220a791b631540e19c8cdaa4abfaa84e7b49fa4d038b2f9909ea05a3d4c4"} Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.587674 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.587655679 podStartE2EDuration="2.587655679s" podCreationTimestamp="2025-10-02 13:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:50.57739094 +0000 UTC m=+1354.683801853" watchObservedRunningTime="2025-10-02 13:11:50.587655679 +0000 UTC m=+1354.694066552" Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.676646 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 13:11:50 crc kubenswrapper[4710]: W1002 13:11:50.683384 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bd1f002_8396_4368_8d3c_432a2590da5c.slice/crio-ac1262ca19679f70dbd81db1e15375f8cc5018bbe276b36b477397704f39d430 WatchSource:0}: Error finding container ac1262ca19679f70dbd81db1e15375f8cc5018bbe276b36b477397704f39d430: Status 404 returned error can't find the container with id ac1262ca19679f70dbd81db1e15375f8cc5018bbe276b36b477397704f39d430 Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.916642 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28bb35ad-6443-4ff0-86f0-358b7fe8ecc3" path="/var/lib/kubelet/pods/28bb35ad-6443-4ff0-86f0-358b7fe8ecc3/volumes" Oct 02 13:11:50 crc kubenswrapper[4710]: I1002 13:11:50.917298 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="381e640f-1570-4d2f-876f-f0a36147712e" path="/var/lib/kubelet/pods/381e640f-1570-4d2f-876f-f0a36147712e/volumes" Oct 02 13:11:51 crc kubenswrapper[4710]: I1002 13:11:51.573135 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8bd1f002-8396-4368-8d3c-432a2590da5c","Type":"ContainerStarted","Data":"6daf6964f265e0027275629399e71c05a311a8cb5c52e73cb6e4a3f8b24297fa"} Oct 02 13:11:51 crc kubenswrapper[4710]: I1002 13:11:51.573443 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8bd1f002-8396-4368-8d3c-432a2590da5c","Type":"ContainerStarted","Data":"a1670fc7d41b0978d60dd2928272afba48ec8e21354a422c244ec68599b2d067"} Oct 02 13:11:51 crc kubenswrapper[4710]: I1002 13:11:51.573453 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8bd1f002-8396-4368-8d3c-432a2590da5c","Type":"ContainerStarted","Data":"ac1262ca19679f70dbd81db1e15375f8cc5018bbe276b36b477397704f39d430"} Oct 02 13:11:51 crc kubenswrapper[4710]: I1002 13:11:51.579089 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"34e05f6e-7f38-4c5f-88fc-31be5ed5a01a","Type":"ContainerStarted","Data":"9d74933592c80d8a54dec61bcfa2c5901bcfd4f00fd7455574e2894153d2d5e4"} Oct 02 13:11:51 crc kubenswrapper[4710]: I1002 13:11:51.598235 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5982142379999997 podStartE2EDuration="2.598214238s" podCreationTimestamp="2025-10-02 13:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:51.596081944 +0000 UTC m=+1355.702492827" watchObservedRunningTime="2025-10-02 13:11:51.598214238 +0000 UTC m=+1355.704625121" Oct 02 13:11:51 crc kubenswrapper[4710]: I1002 13:11:51.616613 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.6165946719999997 podStartE2EDuration="2.616594672s" podCreationTimestamp="2025-10-02 13:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:11:51.613120785 +0000 UTC m=+1355.719531668" watchObservedRunningTime="2025-10-02 13:11:51.616594672 +0000 UTC m=+1355.723005565" Oct 02 13:11:52 crc kubenswrapper[4710]: I1002 13:11:52.530317 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:11:52 crc kubenswrapper[4710]: I1002 13:11:52.530670 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:11:53 crc kubenswrapper[4710]: I1002 13:11:53.962795 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 13:11:53 crc kubenswrapper[4710]: I1002 13:11:53.962864 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 13:11:55 crc kubenswrapper[4710]: I1002 13:11:55.064139 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 13:11:58 crc kubenswrapper[4710]: I1002 13:11:58.962912 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 13:11:58 crc kubenswrapper[4710]: I1002 13:11:58.963652 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 13:11:59 crc kubenswrapper[4710]: I1002 13:11:59.980056 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e55e8319-4c74-4b16-b045-a6f37f4a1ab5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.223:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 13:11:59 crc kubenswrapper[4710]: I1002 13:11:59.980105 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e55e8319-4c74-4b16-b045-a6f37f4a1ab5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.223:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 13:12:00 crc kubenswrapper[4710]: I1002 13:12:00.064998 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 13:12:00 crc kubenswrapper[4710]: I1002 13:12:00.088316 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:12:00 crc kubenswrapper[4710]: I1002 13:12:00.088397 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 13:12:00 crc kubenswrapper[4710]: I1002 13:12:00.095690 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 13:12:00 crc kubenswrapper[4710]: I1002 13:12:00.714317 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 13:12:01 crc kubenswrapper[4710]: I1002 13:12:01.105912 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8bd1f002-8396-4368-8d3c-432a2590da5c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 13:12:01 crc kubenswrapper[4710]: I1002 13:12:01.105910 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8bd1f002-8396-4368-8d3c-432a2590da5c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 13:12:01 crc kubenswrapper[4710]: I1002 13:12:01.713034 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 13:12:08 crc kubenswrapper[4710]: I1002 13:12:08.968124 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 13:12:08 crc kubenswrapper[4710]: I1002 13:12:08.972900 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 13:12:08 crc kubenswrapper[4710]: I1002 13:12:08.974030 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 13:12:09 crc kubenswrapper[4710]: I1002 13:12:09.771416 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 13:12:10 crc kubenswrapper[4710]: I1002 13:12:10.096992 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 13:12:10 crc kubenswrapper[4710]: I1002 13:12:10.097692 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 13:12:10 crc kubenswrapper[4710]: I1002 13:12:10.097714 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 13:12:10 crc kubenswrapper[4710]: I1002 13:12:10.106621 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 13:12:10 crc kubenswrapper[4710]: I1002 13:12:10.780561 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 13:12:10 crc kubenswrapper[4710]: I1002 13:12:10.792088 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 13:12:18 crc kubenswrapper[4710]: I1002 13:12:18.872108 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:12:19 crc kubenswrapper[4710]: I1002 13:12:19.592466 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:12:20 crc kubenswrapper[4710]: I1002 13:12:20.838915 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5dfkb"] Oct 02 13:12:20 crc kubenswrapper[4710]: I1002 13:12:20.842077 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:20 crc kubenswrapper[4710]: I1002 13:12:20.859960 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dfkb"] Oct 02 13:12:20 crc kubenswrapper[4710]: I1002 13:12:20.938569 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-utilities\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:20 crc kubenswrapper[4710]: I1002 13:12:20.938633 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-catalog-content\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:20 crc kubenswrapper[4710]: I1002 13:12:20.938661 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smnzq\" (UniqueName: \"kubernetes.io/projected/c9260374-671a-41b0-8210-d8c4d278495b-kube-api-access-smnzq\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.041085 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-utilities\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.041151 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-catalog-content\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.041180 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smnzq\" (UniqueName: \"kubernetes.io/projected/c9260374-671a-41b0-8210-d8c4d278495b-kube-api-access-smnzq\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.052037 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-utilities\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.052283 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-catalog-content\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.086949 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smnzq\" (UniqueName: \"kubernetes.io/projected/c9260374-671a-41b0-8210-d8c4d278495b-kube-api-access-smnzq\") pod \"redhat-operators-5dfkb\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.166129 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.799351 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dfkb"] Oct 02 13:12:21 crc kubenswrapper[4710]: W1002 13:12:21.813655 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9260374_671a_41b0_8210_d8c4d278495b.slice/crio-8f55e77ebd269b305b77eac45b9770064be818ebdb3c5af022a0ff28415c2906 WatchSource:0}: Error finding container 8f55e77ebd269b305b77eac45b9770064be818ebdb3c5af022a0ff28415c2906: Status 404 returned error can't find the container with id 8f55e77ebd269b305b77eac45b9770064be818ebdb3c5af022a0ff28415c2906 Oct 02 13:12:21 crc kubenswrapper[4710]: I1002 13:12:21.896065 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerStarted","Data":"8f55e77ebd269b305b77eac45b9770064be818ebdb3c5af022a0ff28415c2906"} Oct 02 13:12:22 crc kubenswrapper[4710]: I1002 13:12:22.530179 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:12:22 crc kubenswrapper[4710]: I1002 13:12:22.530730 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:12:22 crc kubenswrapper[4710]: I1002 13:12:22.709835 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="rabbitmq" containerID="cri-o://2a53df3200ea0655a7b8d77a58300f314704d1f704aac912604d3e7e26ffdf3e" gracePeriod=604797 Oct 02 13:12:22 crc kubenswrapper[4710]: I1002 13:12:22.912707 4710 generic.go:334] "Generic (PLEG): container finished" podID="c9260374-671a-41b0-8210-d8c4d278495b" containerID="f542341d24d05eeef5796236558be49881216b55d324ae55847b386a0ae3b051" exitCode=0 Oct 02 13:12:22 crc kubenswrapper[4710]: I1002 13:12:22.924599 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerDied","Data":"f542341d24d05eeef5796236558be49881216b55d324ae55847b386a0ae3b051"} Oct 02 13:12:23 crc kubenswrapper[4710]: I1002 13:12:23.182355 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="rabbitmq" containerID="cri-o://5e0e2863df8edcbcd5a8ff5b36093c5369fa9833fab9d8d8b15a4dda623e23e3" gracePeriod=604797 Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.470653 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.939231 4710 generic.go:334] "Generic (PLEG): container finished" podID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerID="5e0e2863df8edcbcd5a8ff5b36093c5369fa9833fab9d8d8b15a4dda623e23e3" exitCode=0 Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.939339 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7fb5713-6a35-4676-b466-6a43a4037f7b","Type":"ContainerDied","Data":"5e0e2863df8edcbcd5a8ff5b36093c5369fa9833fab9d8d8b15a4dda623e23e3"} Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.942262 4710 generic.go:334] "Generic (PLEG): container finished" podID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerID="2a53df3200ea0655a7b8d77a58300f314704d1f704aac912604d3e7e26ffdf3e" exitCode=0 Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.942293 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"770af87f-9638-4cdb-a10d-2b40d50302ca","Type":"ContainerDied","Data":"2a53df3200ea0655a7b8d77a58300f314704d1f704aac912604d3e7e26ffdf3e"} Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.942314 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"770af87f-9638-4cdb-a10d-2b40d50302ca","Type":"ContainerDied","Data":"d3ce6a9121dbb564bf1407390305275d872323c79d0b0ed1dd25e300b9e61ca1"} Oct 02 13:12:24 crc kubenswrapper[4710]: I1002 13:12:24.942329 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3ce6a9121dbb564bf1407390305275d872323c79d0b0ed1dd25e300b9e61ca1" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.006355 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.129684 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-server-conf\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.129786 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/770af87f-9638-4cdb-a10d-2b40d50302ca-erlang-cookie-secret\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.129831 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.129940 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/770af87f-9638-4cdb-a10d-2b40d50302ca-pod-info\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.129981 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2fjs\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-kube-api-access-m2fjs\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.130004 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-config-data\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.130034 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-confd\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.130071 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-plugins-conf\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.130097 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-erlang-cookie\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.130134 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-plugins\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.130221 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-tls\") pod \"770af87f-9638-4cdb-a10d-2b40d50302ca\" (UID: \"770af87f-9638-4cdb-a10d-2b40d50302ca\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.138211 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.139809 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.145423 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.148319 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.148867 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.150723 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/770af87f-9638-4cdb-a10d-2b40d50302ca-pod-info" (OuterVolumeSpecName: "pod-info") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.152234 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-kube-api-access-m2fjs" (OuterVolumeSpecName: "kube-api-access-m2fjs") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "kube-api-access-m2fjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.193238 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770af87f-9638-4cdb-a10d-2b40d50302ca-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.212163 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.232515 4710 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/770af87f-9638-4cdb-a10d-2b40d50302ca-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.232819 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.232911 4710 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/770af87f-9638-4cdb-a10d-2b40d50302ca-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.232987 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2fjs\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-kube-api-access-m2fjs\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.233056 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.233126 4710 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.233197 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.233261 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.234974 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-config-data" (OuterVolumeSpecName: "config-data") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.237011 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-server-conf" (OuterVolumeSpecName: "server-conf") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.266539 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.328785 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "770af87f-9638-4cdb-a10d-2b40d50302ca" (UID: "770af87f-9638-4cdb-a10d-2b40d50302ca"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.335082 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-plugins-conf\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.335373 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7fb5713-6a35-4676-b466-6a43a4037f7b-erlang-cookie-secret\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.335445 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-server-conf\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.335591 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-config-data\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.335738 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xghww\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-kube-api-access-xghww\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.335843 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-erlang-cookie\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336018 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7fb5713-6a35-4676-b466-6a43a4037f7b-pod-info\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336047 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-confd\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336071 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336089 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-tls\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336245 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-plugins\") pod \"b7fb5713-6a35-4676-b466-6a43a4037f7b\" (UID: \"b7fb5713-6a35-4676-b466-6a43a4037f7b\") " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336728 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336811 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/770af87f-9638-4cdb-a10d-2b40d50302ca-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336829 4710 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/770af87f-9638-4cdb-a10d-2b40d50302ca-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.336838 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.339973 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.340267 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.341416 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.341577 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.347999 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fb5713-6a35-4676-b466-6a43a4037f7b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.364867 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.372478 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-config-data" (OuterVolumeSpecName: "config-data") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.375200 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b7fb5713-6a35-4676-b466-6a43a4037f7b-pod-info" (OuterVolumeSpecName: "pod-info") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.376162 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-kube-api-access-xghww" (OuterVolumeSpecName: "kube-api-access-xghww") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "kube-api-access-xghww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.412366 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-server-conf" (OuterVolumeSpecName: "server-conf") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438440 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xghww\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-kube-api-access-xghww\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438476 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438488 4710 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7fb5713-6a35-4676-b466-6a43a4037f7b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438519 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438531 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438540 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438550 4710 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438559 4710 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7fb5713-6a35-4676-b466-6a43a4037f7b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438568 4710 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.438578 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7fb5713-6a35-4676-b466-6a43a4037f7b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.445969 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b7fb5713-6a35-4676-b466-6a43a4037f7b" (UID: "b7fb5713-6a35-4676-b466-6a43a4037f7b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.460108 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.539923 4710 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7fb5713-6a35-4676-b466-6a43a4037f7b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.539968 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.964739 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.965133 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.965175 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7fb5713-6a35-4676-b466-6a43a4037f7b","Type":"ContainerDied","Data":"1f0998615266bdef8116a38c01758db832eda01855d14331f8f6db1ae9099f7e"} Oct 02 13:12:25 crc kubenswrapper[4710]: I1002 13:12:25.966146 4710 scope.go:117] "RemoveContainer" containerID="5e0e2863df8edcbcd5a8ff5b36093c5369fa9833fab9d8d8b15a4dda623e23e3" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.012414 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.020282 4710 scope.go:117] "RemoveContainer" containerID="6987be090ed6cf0c897ec0e533b9d01229b7de9b3109dc7f672c9f5516cfef82" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.026632 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.050957 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.063044 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.086091 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: E1002 13:12:26.086785 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="rabbitmq" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.087000 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="rabbitmq" Oct 02 13:12:26 crc kubenswrapper[4710]: E1002 13:12:26.087149 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="setup-container" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.087243 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="setup-container" Oct 02 13:12:26 crc kubenswrapper[4710]: E1002 13:12:26.087323 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="rabbitmq" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.087403 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="rabbitmq" Oct 02 13:12:26 crc kubenswrapper[4710]: E1002 13:12:26.087497 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="setup-container" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.087581 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="setup-container" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.087912 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" containerName="rabbitmq" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.088016 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" containerName="rabbitmq" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.089617 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.094765 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.096095 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.096273 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gzd98" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.096403 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.096503 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.097134 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.098912 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.100061 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.110343 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x6wq4" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.111459 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.113132 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.114052 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.114050 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.114690 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.115053 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.115565 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.145862 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.164799 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.164887 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-config-data\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.164929 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5eb635a-718a-4361-a3b6-33d483c1e1b2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.164956 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165011 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jb4t\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-kube-api-access-7jb4t\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165044 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165083 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5eb635a-718a-4361-a3b6-33d483c1e1b2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165130 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165256 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165329 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st8qr\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-kube-api-access-st8qr\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165365 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165430 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165463 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165500 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165531 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165569 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165772 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165809 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165853 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165881 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165920 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.165999 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.200308 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267590 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st8qr\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-kube-api-access-st8qr\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267646 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267702 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267732 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267877 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267903 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267931 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.267991 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268021 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268052 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268074 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268099 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268149 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268199 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268238 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-config-data\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268267 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5eb635a-718a-4361-a3b6-33d483c1e1b2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268292 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268327 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jb4t\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-kube-api-access-7jb4t\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268351 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268383 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5eb635a-718a-4361-a3b6-33d483c1e1b2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268418 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.268440 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.269044 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.271836 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.272469 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-config-data\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.273006 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.273125 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.273298 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.273448 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.273651 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.274441 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.274475 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.274917 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.276228 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5eb635a-718a-4361-a3b6-33d483c1e1b2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.276573 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.278290 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5eb635a-718a-4361-a3b6-33d483c1e1b2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.278664 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.281378 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.277587 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.286216 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.286980 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5eb635a-718a-4361-a3b6-33d483c1e1b2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.287840 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.288422 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st8qr\" (UniqueName: \"kubernetes.io/projected/a5eb635a-718a-4361-a3b6-33d483c1e1b2-kube-api-access-st8qr\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.293124 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jb4t\" (UniqueName: \"kubernetes.io/projected/a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f-kube-api-access-7jb4t\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.315541 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.324199 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a5eb635a-718a-4361-a3b6-33d483c1e1b2\") " pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.435481 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.460684 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.919308 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770af87f-9638-4cdb-a10d-2b40d50302ca" path="/var/lib/kubelet/pods/770af87f-9638-4cdb-a10d-2b40d50302ca/volumes" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.924172 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7fb5713-6a35-4676-b466-6a43a4037f7b" path="/var/lib/kubelet/pods/b7fb5713-6a35-4676-b466-6a43a4037f7b/volumes" Oct 02 13:12:26 crc kubenswrapper[4710]: I1002 13:12:26.981538 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerStarted","Data":"c192fad0aa3ecc9de8b9978f169495111f49b8a3373adf52227e95e82935aea5"} Oct 02 13:12:27 crc kubenswrapper[4710]: I1002 13:12:27.083431 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 13:12:27 crc kubenswrapper[4710]: I1002 13:12:27.090545 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 13:12:27 crc kubenswrapper[4710]: I1002 13:12:27.995666 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f","Type":"ContainerStarted","Data":"c3f4fc04211ea953067eaa75763cc09d8e603332e96dee0ab6d162c07d92f765"} Oct 02 13:12:27 crc kubenswrapper[4710]: I1002 13:12:27.998309 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a5eb635a-718a-4361-a3b6-33d483c1e1b2","Type":"ContainerStarted","Data":"f41d7ffc1939583558ac1ce2af3c4213a401359ea21c4b140d51dedb68f2456f"} Oct 02 13:12:28 crc kubenswrapper[4710]: I1002 13:12:28.001610 4710 generic.go:334] "Generic (PLEG): container finished" podID="c9260374-671a-41b0-8210-d8c4d278495b" containerID="c192fad0aa3ecc9de8b9978f169495111f49b8a3373adf52227e95e82935aea5" exitCode=0 Oct 02 13:12:28 crc kubenswrapper[4710]: I1002 13:12:28.001647 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerDied","Data":"c192fad0aa3ecc9de8b9978f169495111f49b8a3373adf52227e95e82935aea5"} Oct 02 13:12:30 crc kubenswrapper[4710]: I1002 13:12:30.039049 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a5eb635a-718a-4361-a3b6-33d483c1e1b2","Type":"ContainerStarted","Data":"b0960bbaacc3e92b96e6ba5b04258705acbf493ffa097f4cd657e133e943ced4"} Oct 02 13:12:30 crc kubenswrapper[4710]: I1002 13:12:30.042204 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerStarted","Data":"6cb0c0e41cbdc2b0ef9923b55735f52424c85b6978f3aeb14939de427c0c5873"} Oct 02 13:12:30 crc kubenswrapper[4710]: I1002 13:12:30.045303 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f","Type":"ContainerStarted","Data":"13116f9a1a7315a78b7c72ae0a6e51283786420d8053a6fad7127b1e4cde2d30"} Oct 02 13:12:30 crc kubenswrapper[4710]: I1002 13:12:30.131881 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5dfkb" podStartSLOduration=3.530877954 podStartE2EDuration="10.131850081s" podCreationTimestamp="2025-10-02 13:12:20 +0000 UTC" firstStartedPulling="2025-10-02 13:12:22.914072358 +0000 UTC m=+1387.020483241" lastFinishedPulling="2025-10-02 13:12:29.515044445 +0000 UTC m=+1393.621455368" observedRunningTime="2025-10-02 13:12:30.120539195 +0000 UTC m=+1394.226950078" watchObservedRunningTime="2025-10-02 13:12:30.131850081 +0000 UTC m=+1394.238260964" Oct 02 13:12:31 crc kubenswrapper[4710]: I1002 13:12:31.167737 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:31 crc kubenswrapper[4710]: I1002 13:12:31.167804 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:32 crc kubenswrapper[4710]: I1002 13:12:32.212827 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dfkb" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="registry-server" probeResult="failure" output=< Oct 02 13:12:32 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 13:12:32 crc kubenswrapper[4710]: > Oct 02 13:12:41 crc kubenswrapper[4710]: I1002 13:12:41.222708 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:41 crc kubenswrapper[4710]: I1002 13:12:41.302850 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:41 crc kubenswrapper[4710]: I1002 13:12:41.474844 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dfkb"] Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.868128 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b47fd4dcf-42sf7"] Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.870596 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.875253 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896115 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrnmp\" (UniqueName: \"kubernetes.io/projected/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-kube-api-access-xrnmp\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896219 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-config\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896399 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-svc\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896518 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-swift-storage-0\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896571 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-nb\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896699 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-sb\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.896824 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:42 crc kubenswrapper[4710]: I1002 13:12:42.972569 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b47fd4dcf-42sf7"] Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000144 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-svc\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000249 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-swift-storage-0\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000290 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-nb\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000361 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-sb\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000406 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000456 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrnmp\" (UniqueName: \"kubernetes.io/projected/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-kube-api-access-xrnmp\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.000526 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-config\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.001737 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-config\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.002397 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-svc\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.003041 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-swift-storage-0\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.003672 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-nb\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.004320 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-sb\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.004915 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.046423 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrnmp\" (UniqueName: \"kubernetes.io/projected/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-kube-api-access-xrnmp\") pod \"dnsmasq-dns-6b47fd4dcf-42sf7\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.164313 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5dfkb" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="registry-server" containerID="cri-o://6cb0c0e41cbdc2b0ef9923b55735f52424c85b6978f3aeb14939de427c0c5873" gracePeriod=2 Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.263256 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.594186 4710 scope.go:117] "RemoveContainer" containerID="6b868c2abc568ca5d3358f49917fba3e3214d607ef5d3ce34983a6c316e15ab3" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.658863 4710 scope.go:117] "RemoveContainer" containerID="2a53df3200ea0655a7b8d77a58300f314704d1f704aac912604d3e7e26ffdf3e" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.683236 4710 scope.go:117] "RemoveContainer" containerID="d09fdff87746597e8783db624850e59db4ddbebcd80219d59a2c98e7480230a2" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.725139 4710 scope.go:117] "RemoveContainer" containerID="fdee93596cd8050671b0eb9eb2f1293b5198c88057e712355864ad2002099d9f" Oct 02 13:12:43 crc kubenswrapper[4710]: I1002 13:12:43.726333 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b47fd4dcf-42sf7"] Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.182620 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" event={"ID":"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8","Type":"ContainerStarted","Data":"00e1bac2a9a7869795c362fa856e4492c133a886cc36f75248e332f383fb873d"} Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.186132 4710 generic.go:334] "Generic (PLEG): container finished" podID="c9260374-671a-41b0-8210-d8c4d278495b" containerID="6cb0c0e41cbdc2b0ef9923b55735f52424c85b6978f3aeb14939de427c0c5873" exitCode=0 Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.186166 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerDied","Data":"6cb0c0e41cbdc2b0ef9923b55735f52424c85b6978f3aeb14939de427c0c5873"} Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.186208 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dfkb" event={"ID":"c9260374-671a-41b0-8210-d8c4d278495b","Type":"ContainerDied","Data":"8f55e77ebd269b305b77eac45b9770064be818ebdb3c5af022a0ff28415c2906"} Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.186220 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f55e77ebd269b305b77eac45b9770064be818ebdb3c5af022a0ff28415c2906" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.256954 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.335407 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-utilities\") pod \"c9260374-671a-41b0-8210-d8c4d278495b\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.339957 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smnzq\" (UniqueName: \"kubernetes.io/projected/c9260374-671a-41b0-8210-d8c4d278495b-kube-api-access-smnzq\") pod \"c9260374-671a-41b0-8210-d8c4d278495b\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.340247 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-catalog-content\") pod \"c9260374-671a-41b0-8210-d8c4d278495b\" (UID: \"c9260374-671a-41b0-8210-d8c4d278495b\") " Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.337495 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-utilities" (OuterVolumeSpecName: "utilities") pod "c9260374-671a-41b0-8210-d8c4d278495b" (UID: "c9260374-671a-41b0-8210-d8c4d278495b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.373097 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9260374-671a-41b0-8210-d8c4d278495b-kube-api-access-smnzq" (OuterVolumeSpecName: "kube-api-access-smnzq") pod "c9260374-671a-41b0-8210-d8c4d278495b" (UID: "c9260374-671a-41b0-8210-d8c4d278495b"). InnerVolumeSpecName "kube-api-access-smnzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.445681 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9260374-671a-41b0-8210-d8c4d278495b" (UID: "c9260374-671a-41b0-8210-d8c4d278495b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.448420 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.448468 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9260374-671a-41b0-8210-d8c4d278495b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:44 crc kubenswrapper[4710]: I1002 13:12:44.448481 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smnzq\" (UniqueName: \"kubernetes.io/projected/c9260374-671a-41b0-8210-d8c4d278495b-kube-api-access-smnzq\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:45 crc kubenswrapper[4710]: I1002 13:12:45.199713 4710 generic.go:334] "Generic (PLEG): container finished" podID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerID="9ceb81386ab6229df9b3bd474933fc0e9df2256e664c80316071016d9532baab" exitCode=0 Oct 02 13:12:45 crc kubenswrapper[4710]: I1002 13:12:45.199799 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" event={"ID":"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8","Type":"ContainerDied","Data":"9ceb81386ab6229df9b3bd474933fc0e9df2256e664c80316071016d9532baab"} Oct 02 13:12:45 crc kubenswrapper[4710]: I1002 13:12:45.200077 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dfkb" Oct 02 13:12:45 crc kubenswrapper[4710]: I1002 13:12:45.263052 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dfkb"] Oct 02 13:12:45 crc kubenswrapper[4710]: I1002 13:12:45.276971 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5dfkb"] Oct 02 13:12:46 crc kubenswrapper[4710]: I1002 13:12:46.213855 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" event={"ID":"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8","Type":"ContainerStarted","Data":"1f555cef3f3e5bed36663cdf667446b3e2dcb0d0a47e8a0178633ac34985316c"} Oct 02 13:12:46 crc kubenswrapper[4710]: I1002 13:12:46.214472 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:46 crc kubenswrapper[4710]: I1002 13:12:46.248581 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" podStartSLOduration=4.2485581119999996 podStartE2EDuration="4.248558112s" podCreationTimestamp="2025-10-02 13:12:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:12:46.236904758 +0000 UTC m=+1410.343315651" watchObservedRunningTime="2025-10-02 13:12:46.248558112 +0000 UTC m=+1410.354969005" Oct 02 13:12:46 crc kubenswrapper[4710]: I1002 13:12:46.916133 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9260374-671a-41b0-8210-d8c4d278495b" path="/var/lib/kubelet/pods/c9260374-671a-41b0-8210-d8c4d278495b/volumes" Oct 02 13:12:52 crc kubenswrapper[4710]: I1002 13:12:52.530786 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:12:52 crc kubenswrapper[4710]: I1002 13:12:52.532124 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:12:52 crc kubenswrapper[4710]: I1002 13:12:52.532225 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:12:52 crc kubenswrapper[4710]: I1002 13:12:52.533079 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b5183e50d1f99d425b43adb71d4ff8d51ea84c4eb7c8acc094921b9bbeee6bd"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:12:52 crc kubenswrapper[4710]: I1002 13:12:52.533213 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://4b5183e50d1f99d425b43adb71d4ff8d51ea84c4eb7c8acc094921b9bbeee6bd" gracePeriod=600 Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.264970 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.283883 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="4b5183e50d1f99d425b43adb71d4ff8d51ea84c4eb7c8acc094921b9bbeee6bd" exitCode=0 Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.283956 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"4b5183e50d1f99d425b43adb71d4ff8d51ea84c4eb7c8acc094921b9bbeee6bd"} Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.284321 4710 scope.go:117] "RemoveContainer" containerID="51e94e7f2d48ee2b42323dae0e2a0810cf3081a0dd80ed3b534b22c393fd779b" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.332149 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b87b8bb9-9pdwf"] Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.332407 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerName="dnsmasq-dns" containerID="cri-o://15e2a39a6cf9ceab890ac8e1c35cc9fe9369ab8cd4104773e6d068b4285bf489" gracePeriod=10 Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.479157 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dc799465c-hvxxq"] Oct 02 13:12:53 crc kubenswrapper[4710]: E1002 13:12:53.479606 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="registry-server" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.479626 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="registry-server" Oct 02 13:12:53 crc kubenswrapper[4710]: E1002 13:12:53.479645 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="extract-content" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.479653 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="extract-content" Oct 02 13:12:53 crc kubenswrapper[4710]: E1002 13:12:53.479663 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="extract-utilities" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.479670 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="extract-utilities" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.479919 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9260374-671a-41b0-8210-d8c4d278495b" containerName="registry-server" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.481052 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.514519 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc799465c-hvxxq"] Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.670411 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr6rs\" (UniqueName: \"kubernetes.io/projected/b611b238-2861-4371-bbdc-6cfb79039ca4-kube-api-access-jr6rs\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.671288 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-config\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.671368 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-dns-swift-storage-0\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.671407 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-ovsdbserver-sb\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.671431 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-ovsdbserver-nb\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.671463 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-dns-svc\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.671993 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-openstack-edpm-ipam\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.774504 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-openstack-edpm-ipam\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.774994 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr6rs\" (UniqueName: \"kubernetes.io/projected/b611b238-2861-4371-bbdc-6cfb79039ca4-kube-api-access-jr6rs\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.775115 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-config\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.775226 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-dns-swift-storage-0\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.775320 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-ovsdbserver-sb\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.775435 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-ovsdbserver-nb\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.775529 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-dns-svc\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.775552 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-openstack-edpm-ipam\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.776119 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-dns-swift-storage-0\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.776879 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-ovsdbserver-sb\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.777091 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-config\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.777604 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-dns-svc\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.777921 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b611b238-2861-4371-bbdc-6cfb79039ca4-ovsdbserver-nb\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.797661 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr6rs\" (UniqueName: \"kubernetes.io/projected/b611b238-2861-4371-bbdc-6cfb79039ca4-kube-api-access-jr6rs\") pod \"dnsmasq-dns-dc799465c-hvxxq\" (UID: \"b611b238-2861-4371-bbdc-6cfb79039ca4\") " pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:53 crc kubenswrapper[4710]: I1002 13:12:53.814782 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.299999 4710 generic.go:334] "Generic (PLEG): container finished" podID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerID="15e2a39a6cf9ceab890ac8e1c35cc9fe9369ab8cd4104773e6d068b4285bf489" exitCode=0 Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.300085 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" event={"ID":"93d26107-27ce-422c-b17d-ae46b170f8b2","Type":"ContainerDied","Data":"15e2a39a6cf9ceab890ac8e1c35cc9fe9369ab8cd4104773e6d068b4285bf489"} Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.313911 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e"} Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.418564 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc799465c-hvxxq"] Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.659561 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.795660 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-svc\") pod \"93d26107-27ce-422c-b17d-ae46b170f8b2\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.795982 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-nb\") pod \"93d26107-27ce-422c-b17d-ae46b170f8b2\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.796098 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-config\") pod \"93d26107-27ce-422c-b17d-ae46b170f8b2\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.796138 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-swift-storage-0\") pod \"93d26107-27ce-422c-b17d-ae46b170f8b2\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.796218 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xdd6\" (UniqueName: \"kubernetes.io/projected/93d26107-27ce-422c-b17d-ae46b170f8b2-kube-api-access-6xdd6\") pod \"93d26107-27ce-422c-b17d-ae46b170f8b2\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.796334 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-sb\") pod \"93d26107-27ce-422c-b17d-ae46b170f8b2\" (UID: \"93d26107-27ce-422c-b17d-ae46b170f8b2\") " Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.810186 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d26107-27ce-422c-b17d-ae46b170f8b2-kube-api-access-6xdd6" (OuterVolumeSpecName: "kube-api-access-6xdd6") pod "93d26107-27ce-422c-b17d-ae46b170f8b2" (UID: "93d26107-27ce-422c-b17d-ae46b170f8b2"). InnerVolumeSpecName "kube-api-access-6xdd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.861219 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "93d26107-27ce-422c-b17d-ae46b170f8b2" (UID: "93d26107-27ce-422c-b17d-ae46b170f8b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.872841 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-config" (OuterVolumeSpecName: "config") pod "93d26107-27ce-422c-b17d-ae46b170f8b2" (UID: "93d26107-27ce-422c-b17d-ae46b170f8b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.873440 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "93d26107-27ce-422c-b17d-ae46b170f8b2" (UID: "93d26107-27ce-422c-b17d-ae46b170f8b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.880539 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93d26107-27ce-422c-b17d-ae46b170f8b2" (UID: "93d26107-27ce-422c-b17d-ae46b170f8b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.884285 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "93d26107-27ce-422c-b17d-ae46b170f8b2" (UID: "93d26107-27ce-422c-b17d-ae46b170f8b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.899108 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.899153 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.899166 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xdd6\" (UniqueName: \"kubernetes.io/projected/93d26107-27ce-422c-b17d-ae46b170f8b2-kube-api-access-6xdd6\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.899175 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.899184 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:54 crc kubenswrapper[4710]: I1002 13:12:54.899192 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93d26107-27ce-422c-b17d-ae46b170f8b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.334126 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" event={"ID":"93d26107-27ce-422c-b17d-ae46b170f8b2","Type":"ContainerDied","Data":"0f0fb2eb9a93f69d59c2c6d16af450187e8108daa7d09f554bd0c06dee0be8df"} Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.334463 4710 scope.go:117] "RemoveContainer" containerID="15e2a39a6cf9ceab890ac8e1c35cc9fe9369ab8cd4104773e6d068b4285bf489" Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.334574 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b87b8bb9-9pdwf" Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.338587 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" event={"ID":"b611b238-2861-4371-bbdc-6cfb79039ca4","Type":"ContainerStarted","Data":"1ccbe6039d63ba5488d92cbc8c2d67f1e48d661e09674421cae1f1efa5dbaefa"} Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.338618 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" event={"ID":"b611b238-2861-4371-bbdc-6cfb79039ca4","Type":"ContainerStarted","Data":"53635104af9bc2804c8a2c885553ec481be2ca8f3cd9898c60d1a4628ef6606f"} Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.364425 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b87b8bb9-9pdwf"] Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.367262 4710 scope.go:117] "RemoveContainer" containerID="5cdf0f5de9a003e9d34615be5237e879c00c84c25a2037527087099e08f637de" Oct 02 13:12:55 crc kubenswrapper[4710]: I1002 13:12:55.374341 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b87b8bb9-9pdwf"] Oct 02 13:12:56 crc kubenswrapper[4710]: I1002 13:12:56.353522 4710 generic.go:334] "Generic (PLEG): container finished" podID="b611b238-2861-4371-bbdc-6cfb79039ca4" containerID="1ccbe6039d63ba5488d92cbc8c2d67f1e48d661e09674421cae1f1efa5dbaefa" exitCode=0 Oct 02 13:12:56 crc kubenswrapper[4710]: I1002 13:12:56.353649 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" event={"ID":"b611b238-2861-4371-bbdc-6cfb79039ca4","Type":"ContainerDied","Data":"1ccbe6039d63ba5488d92cbc8c2d67f1e48d661e09674421cae1f1efa5dbaefa"} Oct 02 13:12:56 crc kubenswrapper[4710]: I1002 13:12:56.924036 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" path="/var/lib/kubelet/pods/93d26107-27ce-422c-b17d-ae46b170f8b2/volumes" Oct 02 13:12:57 crc kubenswrapper[4710]: I1002 13:12:57.368613 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" event={"ID":"b611b238-2861-4371-bbdc-6cfb79039ca4","Type":"ContainerStarted","Data":"036d78705e330271578b2e978a67fbdb14db8ff9f2e9783e9ca1e970135198da"} Oct 02 13:12:58 crc kubenswrapper[4710]: I1002 13:12:58.376656 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:12:58 crc kubenswrapper[4710]: I1002 13:12:58.400072 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" podStartSLOduration=5.400054368 podStartE2EDuration="5.400054368s" podCreationTimestamp="2025-10-02 13:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:12:58.39656616 +0000 UTC m=+1422.502977043" watchObservedRunningTime="2025-10-02 13:12:58.400054368 +0000 UTC m=+1422.506465251" Oct 02 13:13:01 crc kubenswrapper[4710]: I1002 13:13:01.406099 4710 generic.go:334] "Generic (PLEG): container finished" podID="a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f" containerID="13116f9a1a7315a78b7c72ae0a6e51283786420d8053a6fad7127b1e4cde2d30" exitCode=0 Oct 02 13:13:01 crc kubenswrapper[4710]: I1002 13:13:01.406169 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f","Type":"ContainerDied","Data":"13116f9a1a7315a78b7c72ae0a6e51283786420d8053a6fad7127b1e4cde2d30"} Oct 02 13:13:01 crc kubenswrapper[4710]: I1002 13:13:01.409474 4710 generic.go:334] "Generic (PLEG): container finished" podID="a5eb635a-718a-4361-a3b6-33d483c1e1b2" containerID="b0960bbaacc3e92b96e6ba5b04258705acbf493ffa097f4cd657e133e943ced4" exitCode=0 Oct 02 13:13:01 crc kubenswrapper[4710]: I1002 13:13:01.409507 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a5eb635a-718a-4361-a3b6-33d483c1e1b2","Type":"ContainerDied","Data":"b0960bbaacc3e92b96e6ba5b04258705acbf493ffa097f4cd657e133e943ced4"} Oct 02 13:13:02 crc kubenswrapper[4710]: I1002 13:13:02.434831 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a5eb635a-718a-4361-a3b6-33d483c1e1b2","Type":"ContainerStarted","Data":"db902b37402af7c09ebefab8fb244a618f685fd58714f9ca03e7276272e5a28e"} Oct 02 13:13:02 crc kubenswrapper[4710]: I1002 13:13:02.435389 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 13:13:02 crc kubenswrapper[4710]: I1002 13:13:02.445661 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f","Type":"ContainerStarted","Data":"92f86490c7e0fb9e366394be09ffea62e59c2f0fc66d79ff8f6df452ac950cb0"} Oct 02 13:13:02 crc kubenswrapper[4710]: I1002 13:13:02.446319 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:13:02 crc kubenswrapper[4710]: I1002 13:13:02.472247 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.472229062 podStartE2EDuration="36.472229062s" podCreationTimestamp="2025-10-02 13:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:13:02.466570433 +0000 UTC m=+1426.572981326" watchObservedRunningTime="2025-10-02 13:13:02.472229062 +0000 UTC m=+1426.578639945" Oct 02 13:13:02 crc kubenswrapper[4710]: I1002 13:13:02.505389 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.505371476 podStartE2EDuration="36.505371476s" podCreationTimestamp="2025-10-02 13:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:13:02.502320551 +0000 UTC m=+1426.608731434" watchObservedRunningTime="2025-10-02 13:13:02.505371476 +0000 UTC m=+1426.611782359" Oct 02 13:13:03 crc kubenswrapper[4710]: I1002 13:13:03.817454 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dc799465c-hvxxq" Oct 02 13:13:03 crc kubenswrapper[4710]: I1002 13:13:03.909580 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b47fd4dcf-42sf7"] Oct 02 13:13:03 crc kubenswrapper[4710]: I1002 13:13:03.920965 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerName="dnsmasq-dns" containerID="cri-o://1f555cef3f3e5bed36663cdf667446b3e2dcb0d0a47e8a0178633ac34985316c" gracePeriod=10 Oct 02 13:13:04 crc kubenswrapper[4710]: I1002 13:13:04.466324 4710 generic.go:334] "Generic (PLEG): container finished" podID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerID="1f555cef3f3e5bed36663cdf667446b3e2dcb0d0a47e8a0178633ac34985316c" exitCode=0 Oct 02 13:13:04 crc kubenswrapper[4710]: I1002 13:13:04.466406 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" event={"ID":"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8","Type":"ContainerDied","Data":"1f555cef3f3e5bed36663cdf667446b3e2dcb0d0a47e8a0178633ac34985316c"} Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.107837 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259448 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-sb\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259501 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-nb\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259547 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-config\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259594 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-svc\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259676 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-openstack-edpm-ipam\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259762 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrnmp\" (UniqueName: \"kubernetes.io/projected/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-kube-api-access-xrnmp\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.259854 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-swift-storage-0\") pod \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\" (UID: \"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8\") " Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.268779 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-kube-api-access-xrnmp" (OuterVolumeSpecName: "kube-api-access-xrnmp") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "kube-api-access-xrnmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.324901 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-thwj5"] Oct 02 13:13:05 crc kubenswrapper[4710]: E1002 13:13:05.325449 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerName="init" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.325468 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerName="init" Oct 02 13:13:05 crc kubenswrapper[4710]: E1002 13:13:05.325490 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerName="dnsmasq-dns" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.325498 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerName="dnsmasq-dns" Oct 02 13:13:05 crc kubenswrapper[4710]: E1002 13:13:05.325510 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerName="init" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.325519 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerName="init" Oct 02 13:13:05 crc kubenswrapper[4710]: E1002 13:13:05.325561 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerName="dnsmasq-dns" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.325569 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerName="dnsmasq-dns" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.326000 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" containerName="dnsmasq-dns" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.326019 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d26107-27ce-422c-b17d-ae46b170f8b2" containerName="dnsmasq-dns" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.328052 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.348796 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thwj5"] Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.354083 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-config" (OuterVolumeSpecName: "config") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.372863 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.373379 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrnmp\" (UniqueName: \"kubernetes.io/projected/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-kube-api-access-xrnmp\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.403540 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.420662 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.447391 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.456349 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.456386 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" (UID: "eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.475457 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pglzl\" (UniqueName: \"kubernetes.io/projected/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-kube-api-access-pglzl\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.475523 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-utilities\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.475953 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-catalog-content\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.476153 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.476175 4710 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.476188 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.476199 4710 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.476235 4710 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.479420 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" event={"ID":"eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8","Type":"ContainerDied","Data":"00e1bac2a9a7869795c362fa856e4492c133a886cc36f75248e332f383fb873d"} Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.479475 4710 scope.go:117] "RemoveContainer" containerID="1f555cef3f3e5bed36663cdf667446b3e2dcb0d0a47e8a0178633ac34985316c" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.479528 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b47fd4dcf-42sf7" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.537307 4710 scope.go:117] "RemoveContainer" containerID="9ceb81386ab6229df9b3bd474933fc0e9df2256e664c80316071016d9532baab" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.551477 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b47fd4dcf-42sf7"] Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.570356 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b47fd4dcf-42sf7"] Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.578627 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-catalog-content\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.578820 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pglzl\" (UniqueName: \"kubernetes.io/projected/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-kube-api-access-pglzl\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.578857 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-utilities\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.580289 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-catalog-content\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.580287 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-utilities\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.596188 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pglzl\" (UniqueName: \"kubernetes.io/projected/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-kube-api-access-pglzl\") pod \"redhat-marketplace-thwj5\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:05 crc kubenswrapper[4710]: I1002 13:13:05.828884 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:06 crc kubenswrapper[4710]: I1002 13:13:06.225181 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thwj5"] Oct 02 13:13:06 crc kubenswrapper[4710]: I1002 13:13:06.489039 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thwj5" event={"ID":"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f","Type":"ContainerStarted","Data":"f47d63d2b6f03c4f4c42b739344303b1b5f741f10a70c83c6fdf3d1bafe9e48b"} Oct 02 13:13:06 crc kubenswrapper[4710]: I1002 13:13:06.917287 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8" path="/var/lib/kubelet/pods/eb1c4dff-6c8b-44d2-8d00-77b28cc75fd8/volumes" Oct 02 13:13:07 crc kubenswrapper[4710]: I1002 13:13:07.507112 4710 generic.go:334] "Generic (PLEG): container finished" podID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerID="04b8173ad405c4c16c4c56cedc6a05ac49d8b55e18a0bc80c07979a99a92efc8" exitCode=0 Oct 02 13:13:07 crc kubenswrapper[4710]: I1002 13:13:07.507158 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thwj5" event={"ID":"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f","Type":"ContainerDied","Data":"04b8173ad405c4c16c4c56cedc6a05ac49d8b55e18a0bc80c07979a99a92efc8"} Oct 02 13:13:11 crc kubenswrapper[4710]: I1002 13:13:11.549566 4710 generic.go:334] "Generic (PLEG): container finished" podID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerID="445f94a71b6aa4e68738998a731ad439e88dcc246082ad20d3a4e05ed2d58611" exitCode=0 Oct 02 13:13:11 crc kubenswrapper[4710]: I1002 13:13:11.550155 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thwj5" event={"ID":"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f","Type":"ContainerDied","Data":"445f94a71b6aa4e68738998a731ad439e88dcc246082ad20d3a4e05ed2d58611"} Oct 02 13:13:12 crc kubenswrapper[4710]: I1002 13:13:12.564000 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thwj5" event={"ID":"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f","Type":"ContainerStarted","Data":"997899622a860d5f960fb6e5d379ff36ca9a6df084ff070c9dbe9c968e6fd746"} Oct 02 13:13:12 crc kubenswrapper[4710]: I1002 13:13:12.588017 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-thwj5" podStartSLOduration=2.846114611 podStartE2EDuration="7.587998987s" podCreationTimestamp="2025-10-02 13:13:05 +0000 UTC" firstStartedPulling="2025-10-02 13:13:07.509398126 +0000 UTC m=+1431.615809019" lastFinishedPulling="2025-10-02 13:13:12.251282512 +0000 UTC m=+1436.357693395" observedRunningTime="2025-10-02 13:13:12.582724938 +0000 UTC m=+1436.689135821" watchObservedRunningTime="2025-10-02 13:13:12.587998987 +0000 UTC m=+1436.694409870" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.138581 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5hdq5"] Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.147992 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.148139 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5hdq5"] Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.260157 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-catalog-content\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.260715 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-utilities\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.260814 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpcld\" (UniqueName: \"kubernetes.io/projected/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-kube-api-access-wpcld\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.362690 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpcld\" (UniqueName: \"kubernetes.io/projected/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-kube-api-access-wpcld\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.362998 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-catalog-content\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.363040 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-utilities\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.363593 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-utilities\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.363667 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-catalog-content\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.382718 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpcld\" (UniqueName: \"kubernetes.io/projected/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-kube-api-access-wpcld\") pod \"community-operators-5hdq5\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:14 crc kubenswrapper[4710]: I1002 13:13:14.535696 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.033487 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5hdq5"] Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.612929 4710 generic.go:334] "Generic (PLEG): container finished" podID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerID="b0a320ef5f1463bbd3d1ea9a2ee76e4841c8fbe56f23cc2d5d75c99f33c100b4" exitCode=0 Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.613155 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerDied","Data":"b0a320ef5f1463bbd3d1ea9a2ee76e4841c8fbe56f23cc2d5d75c99f33c100b4"} Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.613223 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerStarted","Data":"0763a6705284653207acc1fe4bb3b8cb79bb2debe6b725995eeaac34133ae676"} Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.830134 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.830502 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:15 crc kubenswrapper[4710]: I1002 13:13:15.885630 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:16 crc kubenswrapper[4710]: I1002 13:13:16.439303 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.227:5671: connect: connection refused" Oct 02 13:13:16 crc kubenswrapper[4710]: I1002 13:13:16.463820 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a5eb635a-718a-4361-a3b6-33d483c1e1b2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.228:5671: connect: connection refused" Oct 02 13:13:21 crc kubenswrapper[4710]: I1002 13:13:21.685979 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerStarted","Data":"941c12d268bf5efa5e95d922af3a88c88997beabb853c07181016050eeeb99c0"} Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.361812 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l"] Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.363788 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.366682 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.366842 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.367336 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.367559 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.372458 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l"] Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.426119 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f7f5\" (UniqueName: \"kubernetes.io/projected/3086d3aa-5da8-4765-87dc-f7a122984390-kube-api-access-2f7f5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.426166 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.426216 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.426327 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.528683 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f7f5\" (UniqueName: \"kubernetes.io/projected/3086d3aa-5da8-4765-87dc-f7a122984390-kube-api-access-2f7f5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.528777 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.528827 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.528856 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.534769 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.534869 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.534999 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.547812 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f7f5\" (UniqueName: \"kubernetes.io/projected/3086d3aa-5da8-4765-87dc-f7a122984390-kube-api-access-2f7f5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.684953 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.710131 4710 generic.go:334] "Generic (PLEG): container finished" podID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerID="941c12d268bf5efa5e95d922af3a88c88997beabb853c07181016050eeeb99c0" exitCode=0 Oct 02 13:13:22 crc kubenswrapper[4710]: I1002 13:13:22.710181 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerDied","Data":"941c12d268bf5efa5e95d922af3a88c88997beabb853c07181016050eeeb99c0"} Oct 02 13:13:23 crc kubenswrapper[4710]: W1002 13:13:23.342533 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3086d3aa_5da8_4765_87dc_f7a122984390.slice/crio-7a0fa400dfceaebbdf58913e40fedcf8257064e29c0873902a5c9c0a8a2e5c59 WatchSource:0}: Error finding container 7a0fa400dfceaebbdf58913e40fedcf8257064e29c0873902a5c9c0a8a2e5c59: Status 404 returned error can't find the container with id 7a0fa400dfceaebbdf58913e40fedcf8257064e29c0873902a5c9c0a8a2e5c59 Oct 02 13:13:23 crc kubenswrapper[4710]: I1002 13:13:23.353860 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l"] Oct 02 13:13:23 crc kubenswrapper[4710]: I1002 13:13:23.739953 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" event={"ID":"3086d3aa-5da8-4765-87dc-f7a122984390","Type":"ContainerStarted","Data":"7a0fa400dfceaebbdf58913e40fedcf8257064e29c0873902a5c9c0a8a2e5c59"} Oct 02 13:13:24 crc kubenswrapper[4710]: I1002 13:13:24.757988 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerStarted","Data":"3fe502bfd9580aed690c14491d8e9ff32aaac8d243c1a5326c385eeed35d25e8"} Oct 02 13:13:24 crc kubenswrapper[4710]: I1002 13:13:24.803875 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5hdq5" podStartSLOduration=2.700182266 podStartE2EDuration="10.803850361s" podCreationTimestamp="2025-10-02 13:13:14 +0000 UTC" firstStartedPulling="2025-10-02 13:13:15.614256881 +0000 UTC m=+1439.720667764" lastFinishedPulling="2025-10-02 13:13:23.717924966 +0000 UTC m=+1447.824335859" observedRunningTime="2025-10-02 13:13:24.790499633 +0000 UTC m=+1448.896910506" watchObservedRunningTime="2025-10-02 13:13:24.803850361 +0000 UTC m=+1448.910261244" Oct 02 13:13:25 crc kubenswrapper[4710]: I1002 13:13:25.877315 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:25 crc kubenswrapper[4710]: I1002 13:13:25.937284 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-thwj5"] Oct 02 13:13:26 crc kubenswrapper[4710]: I1002 13:13:26.438182 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 13:13:26 crc kubenswrapper[4710]: I1002 13:13:26.464696 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 13:13:26 crc kubenswrapper[4710]: I1002 13:13:26.783635 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-thwj5" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="registry-server" containerID="cri-o://997899622a860d5f960fb6e5d379ff36ca9a6df084ff070c9dbe9c968e6fd746" gracePeriod=2 Oct 02 13:13:28 crc kubenswrapper[4710]: I1002 13:13:28.807666 4710 generic.go:334] "Generic (PLEG): container finished" podID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerID="997899622a860d5f960fb6e5d379ff36ca9a6df084ff070c9dbe9c968e6fd746" exitCode=0 Oct 02 13:13:28 crc kubenswrapper[4710]: I1002 13:13:28.807726 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thwj5" event={"ID":"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f","Type":"ContainerDied","Data":"997899622a860d5f960fb6e5d379ff36ca9a6df084ff070c9dbe9c968e6fd746"} Oct 02 13:13:31 crc kubenswrapper[4710]: I1002 13:13:31.963525 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.050614 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-utilities\") pod \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.050999 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-catalog-content\") pod \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.051271 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pglzl\" (UniqueName: \"kubernetes.io/projected/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-kube-api-access-pglzl\") pod \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\" (UID: \"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f\") " Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.051309 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-utilities" (OuterVolumeSpecName: "utilities") pod "068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" (UID: "068ba7b4-a0f7-4c2f-b7be-d62851fccc7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.051855 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.069751 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-kube-api-access-pglzl" (OuterVolumeSpecName: "kube-api-access-pglzl") pod "068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" (UID: "068ba7b4-a0f7-4c2f-b7be-d62851fccc7f"). InnerVolumeSpecName "kube-api-access-pglzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.120555 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" (UID: "068ba7b4-a0f7-4c2f-b7be-d62851fccc7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.153781 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pglzl\" (UniqueName: \"kubernetes.io/projected/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-kube-api-access-pglzl\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.153815 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.858402 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thwj5" event={"ID":"068ba7b4-a0f7-4c2f-b7be-d62851fccc7f","Type":"ContainerDied","Data":"f47d63d2b6f03c4f4c42b739344303b1b5f741f10a70c83c6fdf3d1bafe9e48b"} Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.858467 4710 scope.go:117] "RemoveContainer" containerID="997899622a860d5f960fb6e5d379ff36ca9a6df084ff070c9dbe9c968e6fd746" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.858556 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thwj5" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.891671 4710 scope.go:117] "RemoveContainer" containerID="445f94a71b6aa4e68738998a731ad439e88dcc246082ad20d3a4e05ed2d58611" Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.902384 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-thwj5"] Oct 02 13:13:32 crc kubenswrapper[4710]: I1002 13:13:32.917768 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-thwj5"] Oct 02 13:13:33 crc kubenswrapper[4710]: I1002 13:13:33.251552 4710 scope.go:117] "RemoveContainer" containerID="04b8173ad405c4c16c4c56cedc6a05ac49d8b55e18a0bc80c07979a99a92efc8" Oct 02 13:13:34 crc kubenswrapper[4710]: I1002 13:13:34.536566 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:34 crc kubenswrapper[4710]: I1002 13:13:34.536949 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:34 crc kubenswrapper[4710]: I1002 13:13:34.587899 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:34 crc kubenswrapper[4710]: I1002 13:13:34.924965 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" path="/var/lib/kubelet/pods/068ba7b4-a0f7-4c2f-b7be-d62851fccc7f/volumes" Oct 02 13:13:34 crc kubenswrapper[4710]: I1002 13:13:34.952418 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:35 crc kubenswrapper[4710]: I1002 13:13:35.200626 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5hdq5"] Oct 02 13:13:36 crc kubenswrapper[4710]: I1002 13:13:36.900732 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5hdq5" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="registry-server" containerID="cri-o://3fe502bfd9580aed690c14491d8e9ff32aaac8d243c1a5326c385eeed35d25e8" gracePeriod=2 Oct 02 13:13:38 crc kubenswrapper[4710]: I1002 13:13:38.921955 4710 generic.go:334] "Generic (PLEG): container finished" podID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerID="3fe502bfd9580aed690c14491d8e9ff32aaac8d243c1a5326c385eeed35d25e8" exitCode=0 Oct 02 13:13:38 crc kubenswrapper[4710]: I1002 13:13:38.922025 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerDied","Data":"3fe502bfd9580aed690c14491d8e9ff32aaac8d243c1a5326c385eeed35d25e8"} Oct 02 13:13:41 crc kubenswrapper[4710]: E1002 13:13:41.673804 4710 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 02 13:13:41 crc kubenswrapper[4710]: E1002 13:13:41.674600 4710 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 02 13:13:41 crc kubenswrapper[4710]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 02 13:13:41 crc kubenswrapper[4710]: - hosts: all Oct 02 13:13:41 crc kubenswrapper[4710]: strategy: linear Oct 02 13:13:41 crc kubenswrapper[4710]: tasks: Oct 02 13:13:41 crc kubenswrapper[4710]: - name: Enable podified-repos Oct 02 13:13:41 crc kubenswrapper[4710]: become: true Oct 02 13:13:41 crc kubenswrapper[4710]: ansible.builtin.shell: | Oct 02 13:13:41 crc kubenswrapper[4710]: set -euxo pipefail Oct 02 13:13:41 crc kubenswrapper[4710]: pushd /var/tmp Oct 02 13:13:41 crc kubenswrapper[4710]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Oct 02 13:13:41 crc kubenswrapper[4710]: pushd repo-setup-main Oct 02 13:13:41 crc kubenswrapper[4710]: python3 -m venv ./venv Oct 02 13:13:41 crc kubenswrapper[4710]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Oct 02 13:13:41 crc kubenswrapper[4710]: ./venv/bin/repo-setup current-podified -b antelope Oct 02 13:13:41 crc kubenswrapper[4710]: popd Oct 02 13:13:41 crc kubenswrapper[4710]: rm -rf repo-setup-main Oct 02 13:13:41 crc kubenswrapper[4710]: Oct 02 13:13:41 crc kubenswrapper[4710]: Oct 02 13:13:41 crc kubenswrapper[4710]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 02 13:13:41 crc kubenswrapper[4710]: edpm_override_hosts: openstack-edpm-ipam Oct 02 13:13:41 crc kubenswrapper[4710]: edpm_service_type: repo-setup Oct 02 13:13:41 crc kubenswrapper[4710]: Oct 02 13:13:41 crc kubenswrapper[4710]: Oct 02 13:13:41 crc kubenswrapper[4710]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2f7f5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l_openstack(3086d3aa-5da8-4765-87dc-f7a122984390): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 02 13:13:41 crc kubenswrapper[4710]: > logger="UnhandledError" Oct 02 13:13:41 crc kubenswrapper[4710]: E1002 13:13:41.675762 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" podUID="3086d3aa-5da8-4765-87dc-f7a122984390" Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.919963 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.953604 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hdq5" event={"ID":"2c9f9ea9-de09-4d16-b572-2edf1944f1e8","Type":"ContainerDied","Data":"0763a6705284653207acc1fe4bb3b8cb79bb2debe6b725995eeaac34133ae676"} Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.953662 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hdq5" Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.953667 4710 scope.go:117] "RemoveContainer" containerID="3fe502bfd9580aed690c14491d8e9ff32aaac8d243c1a5326c385eeed35d25e8" Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.954827 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpcld\" (UniqueName: \"kubernetes.io/projected/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-kube-api-access-wpcld\") pod \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.954880 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-utilities\") pod \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.954963 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-catalog-content\") pod \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\" (UID: \"2c9f9ea9-de09-4d16-b572-2edf1944f1e8\") " Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.955839 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-utilities" (OuterVolumeSpecName: "utilities") pod "2c9f9ea9-de09-4d16-b572-2edf1944f1e8" (UID: "2c9f9ea9-de09-4d16-b572-2edf1944f1e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:13:41 crc kubenswrapper[4710]: E1002 13:13:41.958666 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" podUID="3086d3aa-5da8-4765-87dc-f7a122984390" Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.967908 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-kube-api-access-wpcld" (OuterVolumeSpecName: "kube-api-access-wpcld") pod "2c9f9ea9-de09-4d16-b572-2edf1944f1e8" (UID: "2c9f9ea9-de09-4d16-b572-2edf1944f1e8"). InnerVolumeSpecName "kube-api-access-wpcld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:13:41 crc kubenswrapper[4710]: I1002 13:13:41.991294 4710 scope.go:117] "RemoveContainer" containerID="941c12d268bf5efa5e95d922af3a88c88997beabb853c07181016050eeeb99c0" Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.040380 4710 scope.go:117] "RemoveContainer" containerID="b0a320ef5f1463bbd3d1ea9a2ee76e4841c8fbe56f23cc2d5d75c99f33c100b4" Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.041477 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c9f9ea9-de09-4d16-b572-2edf1944f1e8" (UID: "2c9f9ea9-de09-4d16-b572-2edf1944f1e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.057119 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpcld\" (UniqueName: \"kubernetes.io/projected/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-kube-api-access-wpcld\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.057161 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.057171 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9f9ea9-de09-4d16-b572-2edf1944f1e8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.288595 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5hdq5"] Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.298157 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5hdq5"] Oct 02 13:13:42 crc kubenswrapper[4710]: I1002 13:13:42.914841 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" path="/var/lib/kubelet/pods/2c9f9ea9-de09-4d16-b572-2edf1944f1e8/volumes" Oct 02 13:13:44 crc kubenswrapper[4710]: I1002 13:13:44.007898 4710 scope.go:117] "RemoveContainer" containerID="ac46934e07d5ee7ef244faaf97821d55a665e7fd407e28f24aa9261894318bb4" Oct 02 13:13:44 crc kubenswrapper[4710]: I1002 13:13:44.041701 4710 scope.go:117] "RemoveContainer" containerID="d83566781e5002da12ac95f0b02ae2974a379a842b8ca9ba535984745afbb884" Oct 02 13:13:44 crc kubenswrapper[4710]: I1002 13:13:44.081317 4710 scope.go:117] "RemoveContainer" containerID="56c067acd68f3ffaa2c02e6108b6721175d00d25a2440ada8b051717cc8a236a" Oct 02 13:13:57 crc kubenswrapper[4710]: I1002 13:13:57.111557 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" event={"ID":"3086d3aa-5da8-4765-87dc-f7a122984390","Type":"ContainerStarted","Data":"88cfe56ff07110108fc7b5746384b0bfee0e7aa1d508dda09c689f143a5846dd"} Oct 02 13:13:57 crc kubenswrapper[4710]: I1002 13:13:57.133258 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" podStartSLOduration=1.9976092109999999 podStartE2EDuration="35.133241969s" podCreationTimestamp="2025-10-02 13:13:22 +0000 UTC" firstStartedPulling="2025-10-02 13:13:23.345156726 +0000 UTC m=+1447.451567619" lastFinishedPulling="2025-10-02 13:13:56.480789494 +0000 UTC m=+1480.587200377" observedRunningTime="2025-10-02 13:13:57.128148134 +0000 UTC m=+1481.234559017" watchObservedRunningTime="2025-10-02 13:13:57.133241969 +0000 UTC m=+1481.239652852" Oct 02 13:14:09 crc kubenswrapper[4710]: I1002 13:14:09.248231 4710 generic.go:334] "Generic (PLEG): container finished" podID="3086d3aa-5da8-4765-87dc-f7a122984390" containerID="88cfe56ff07110108fc7b5746384b0bfee0e7aa1d508dda09c689f143a5846dd" exitCode=0 Oct 02 13:14:09 crc kubenswrapper[4710]: I1002 13:14:09.248310 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" event={"ID":"3086d3aa-5da8-4765-87dc-f7a122984390","Type":"ContainerDied","Data":"88cfe56ff07110108fc7b5746384b0bfee0e7aa1d508dda09c689f143a5846dd"} Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.684607 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.745543 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-ssh-key\") pod \"3086d3aa-5da8-4765-87dc-f7a122984390\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.745719 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f7f5\" (UniqueName: \"kubernetes.io/projected/3086d3aa-5da8-4765-87dc-f7a122984390-kube-api-access-2f7f5\") pod \"3086d3aa-5da8-4765-87dc-f7a122984390\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.745870 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-repo-setup-combined-ca-bundle\") pod \"3086d3aa-5da8-4765-87dc-f7a122984390\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.745902 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-inventory\") pod \"3086d3aa-5da8-4765-87dc-f7a122984390\" (UID: \"3086d3aa-5da8-4765-87dc-f7a122984390\") " Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.750913 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3086d3aa-5da8-4765-87dc-f7a122984390" (UID: "3086d3aa-5da8-4765-87dc-f7a122984390"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.753323 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3086d3aa-5da8-4765-87dc-f7a122984390-kube-api-access-2f7f5" (OuterVolumeSpecName: "kube-api-access-2f7f5") pod "3086d3aa-5da8-4765-87dc-f7a122984390" (UID: "3086d3aa-5da8-4765-87dc-f7a122984390"). InnerVolumeSpecName "kube-api-access-2f7f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.774680 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-inventory" (OuterVolumeSpecName: "inventory") pod "3086d3aa-5da8-4765-87dc-f7a122984390" (UID: "3086d3aa-5da8-4765-87dc-f7a122984390"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.777621 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3086d3aa-5da8-4765-87dc-f7a122984390" (UID: "3086d3aa-5da8-4765-87dc-f7a122984390"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.848239 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.848277 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f7f5\" (UniqueName: \"kubernetes.io/projected/3086d3aa-5da8-4765-87dc-f7a122984390-kube-api-access-2f7f5\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.848289 4710 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:10 crc kubenswrapper[4710]: I1002 13:14:10.848297 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3086d3aa-5da8-4765-87dc-f7a122984390-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.269115 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" event={"ID":"3086d3aa-5da8-4765-87dc-f7a122984390","Type":"ContainerDied","Data":"7a0fa400dfceaebbdf58913e40fedcf8257064e29c0873902a5c9c0a8a2e5c59"} Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.269930 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a0fa400dfceaebbdf58913e40fedcf8257064e29c0873902a5c9c0a8a2e5c59" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.270047 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.366300 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846"] Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367657 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="extract-content" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367690 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="extract-content" Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367731 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="extract-utilities" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367759 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="extract-utilities" Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367783 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="extract-utilities" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367792 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="extract-utilities" Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367824 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3086d3aa-5da8-4765-87dc-f7a122984390" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367834 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3086d3aa-5da8-4765-87dc-f7a122984390" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367851 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="registry-server" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367860 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="registry-server" Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367884 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="registry-server" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367892 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="registry-server" Oct 02 13:14:11 crc kubenswrapper[4710]: E1002 13:14:11.367933 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="extract-content" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.367940 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="extract-content" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.368465 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9f9ea9-de09-4d16-b572-2edf1944f1e8" containerName="registry-server" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.368508 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="068ba7b4-a0f7-4c2f-b7be-d62851fccc7f" containerName="registry-server" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.368548 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3086d3aa-5da8-4765-87dc-f7a122984390" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.370063 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.375465 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.375884 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.377801 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.379578 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.405705 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846"] Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.460636 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.461134 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtgbn\" (UniqueName: \"kubernetes.io/projected/c8c51310-5db6-40e1-b97d-b0460629caed-kube-api-access-dtgbn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.461270 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.563523 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.563584 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtgbn\" (UniqueName: \"kubernetes.io/projected/c8c51310-5db6-40e1-b97d-b0460629caed-kube-api-access-dtgbn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.563676 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.568636 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.569256 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.580660 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtgbn\" (UniqueName: \"kubernetes.io/projected/c8c51310-5db6-40e1-b97d-b0460629caed-kube-api-access-dtgbn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rs846\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:11 crc kubenswrapper[4710]: I1002 13:14:11.696992 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:12 crc kubenswrapper[4710]: I1002 13:14:12.203419 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846"] Oct 02 13:14:12 crc kubenswrapper[4710]: I1002 13:14:12.278775 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" event={"ID":"c8c51310-5db6-40e1-b97d-b0460629caed","Type":"ContainerStarted","Data":"d81c1ff1d6f71651bbb5ac27aa0e6816d968f85eb70a1218a01e551ab92bfdec"} Oct 02 13:14:14 crc kubenswrapper[4710]: I1002 13:14:14.297908 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" event={"ID":"c8c51310-5db6-40e1-b97d-b0460629caed","Type":"ContainerStarted","Data":"0a428159648bce0e19153e3591ff40c0d0ad96a33fff5e240a5ec25e7645ff27"} Oct 02 13:14:14 crc kubenswrapper[4710]: I1002 13:14:14.321596 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" podStartSLOduration=2.458080503 podStartE2EDuration="3.321578219s" podCreationTimestamp="2025-10-02 13:14:11 +0000 UTC" firstStartedPulling="2025-10-02 13:14:12.203634681 +0000 UTC m=+1496.310045564" lastFinishedPulling="2025-10-02 13:14:13.067132397 +0000 UTC m=+1497.173543280" observedRunningTime="2025-10-02 13:14:14.317187731 +0000 UTC m=+1498.423598614" watchObservedRunningTime="2025-10-02 13:14:14.321578219 +0000 UTC m=+1498.427989102" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.055737 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d5bqq"] Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.058770 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.076031 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d5bqq"] Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.136355 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-catalog-content\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.136810 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrc6c\" (UniqueName: \"kubernetes.io/projected/87b2eeb0-fdbf-4acb-9206-60e161ad6646-kube-api-access-nrc6c\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.136973 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-utilities\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.238568 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-catalog-content\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.238632 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrc6c\" (UniqueName: \"kubernetes.io/projected/87b2eeb0-fdbf-4acb-9206-60e161ad6646-kube-api-access-nrc6c\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.238699 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-utilities\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.239274 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-utilities\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.239450 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-catalog-content\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.257408 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrc6c\" (UniqueName: \"kubernetes.io/projected/87b2eeb0-fdbf-4acb-9206-60e161ad6646-kube-api-access-nrc6c\") pod \"certified-operators-d5bqq\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.383559 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:15 crc kubenswrapper[4710]: I1002 13:14:15.832467 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d5bqq"] Oct 02 13:14:15 crc kubenswrapper[4710]: W1002 13:14:15.836048 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87b2eeb0_fdbf_4acb_9206_60e161ad6646.slice/crio-5a65c03ee42e0a12e57477e7e63069253071f53d99b57798f8ba38fae2b053cf WatchSource:0}: Error finding container 5a65c03ee42e0a12e57477e7e63069253071f53d99b57798f8ba38fae2b053cf: Status 404 returned error can't find the container with id 5a65c03ee42e0a12e57477e7e63069253071f53d99b57798f8ba38fae2b053cf Oct 02 13:14:16 crc kubenswrapper[4710]: I1002 13:14:16.319345 4710 generic.go:334] "Generic (PLEG): container finished" podID="c8c51310-5db6-40e1-b97d-b0460629caed" containerID="0a428159648bce0e19153e3591ff40c0d0ad96a33fff5e240a5ec25e7645ff27" exitCode=0 Oct 02 13:14:16 crc kubenswrapper[4710]: I1002 13:14:16.319424 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" event={"ID":"c8c51310-5db6-40e1-b97d-b0460629caed","Type":"ContainerDied","Data":"0a428159648bce0e19153e3591ff40c0d0ad96a33fff5e240a5ec25e7645ff27"} Oct 02 13:14:16 crc kubenswrapper[4710]: I1002 13:14:16.321426 4710 generic.go:334] "Generic (PLEG): container finished" podID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerID="4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9" exitCode=0 Oct 02 13:14:16 crc kubenswrapper[4710]: I1002 13:14:16.321465 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5bqq" event={"ID":"87b2eeb0-fdbf-4acb-9206-60e161ad6646","Type":"ContainerDied","Data":"4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9"} Oct 02 13:14:16 crc kubenswrapper[4710]: I1002 13:14:16.321517 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5bqq" event={"ID":"87b2eeb0-fdbf-4acb-9206-60e161ad6646","Type":"ContainerStarted","Data":"5a65c03ee42e0a12e57477e7e63069253071f53d99b57798f8ba38fae2b053cf"} Oct 02 13:14:17 crc kubenswrapper[4710]: I1002 13:14:17.893542 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.003190 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-ssh-key\") pod \"c8c51310-5db6-40e1-b97d-b0460629caed\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.003258 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-inventory\") pod \"c8c51310-5db6-40e1-b97d-b0460629caed\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.003288 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtgbn\" (UniqueName: \"kubernetes.io/projected/c8c51310-5db6-40e1-b97d-b0460629caed-kube-api-access-dtgbn\") pod \"c8c51310-5db6-40e1-b97d-b0460629caed\" (UID: \"c8c51310-5db6-40e1-b97d-b0460629caed\") " Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.028847 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c51310-5db6-40e1-b97d-b0460629caed-kube-api-access-dtgbn" (OuterVolumeSpecName: "kube-api-access-dtgbn") pod "c8c51310-5db6-40e1-b97d-b0460629caed" (UID: "c8c51310-5db6-40e1-b97d-b0460629caed"). InnerVolumeSpecName "kube-api-access-dtgbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.039104 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c8c51310-5db6-40e1-b97d-b0460629caed" (UID: "c8c51310-5db6-40e1-b97d-b0460629caed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.041560 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-inventory" (OuterVolumeSpecName: "inventory") pod "c8c51310-5db6-40e1-b97d-b0460629caed" (UID: "c8c51310-5db6-40e1-b97d-b0460629caed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.105642 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.105685 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c51310-5db6-40e1-b97d-b0460629caed-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.105700 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtgbn\" (UniqueName: \"kubernetes.io/projected/c8c51310-5db6-40e1-b97d-b0460629caed-kube-api-access-dtgbn\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.346708 4710 generic.go:334] "Generic (PLEG): container finished" podID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerID="87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd" exitCode=0 Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.346864 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5bqq" event={"ID":"87b2eeb0-fdbf-4acb-9206-60e161ad6646","Type":"ContainerDied","Data":"87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd"} Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.350273 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.351109 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" event={"ID":"c8c51310-5db6-40e1-b97d-b0460629caed","Type":"ContainerDied","Data":"d81c1ff1d6f71651bbb5ac27aa0e6816d968f85eb70a1218a01e551ab92bfdec"} Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.351153 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d81c1ff1d6f71651bbb5ac27aa0e6816d968f85eb70a1218a01e551ab92bfdec" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.351232 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rs846" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.421393 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn"] Oct 02 13:14:18 crc kubenswrapper[4710]: E1002 13:14:18.422113 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c51310-5db6-40e1-b97d-b0460629caed" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.422136 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c51310-5db6-40e1-b97d-b0460629caed" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.422355 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c51310-5db6-40e1-b97d-b0460629caed" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.423781 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.427603 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.427787 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.427893 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.429493 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.438293 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn"] Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.512490 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxxj6\" (UniqueName: \"kubernetes.io/projected/65b945dd-164b-45ec-9262-f5d7db0d15dc-kube-api-access-dxxj6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.512626 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.512734 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.512863 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.616635 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.616705 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.616788 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxxj6\" (UniqueName: \"kubernetes.io/projected/65b945dd-164b-45ec-9262-f5d7db0d15dc-kube-api-access-dxxj6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.616843 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.627711 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.628498 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.632644 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.642327 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxxj6\" (UniqueName: \"kubernetes.io/projected/65b945dd-164b-45ec-9262-f5d7db0d15dc-kube-api-access-dxxj6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:18 crc kubenswrapper[4710]: I1002 13:14:18.748007 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:14:19 crc kubenswrapper[4710]: I1002 13:14:19.258776 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn"] Oct 02 13:14:19 crc kubenswrapper[4710]: W1002 13:14:19.259919 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65b945dd_164b_45ec_9262_f5d7db0d15dc.slice/crio-e940c6b2782fc788f8ddc67cf00f2921c1a6b3b945e2517bdb3643acb0709b9d WatchSource:0}: Error finding container e940c6b2782fc788f8ddc67cf00f2921c1a6b3b945e2517bdb3643acb0709b9d: Status 404 returned error can't find the container with id e940c6b2782fc788f8ddc67cf00f2921c1a6b3b945e2517bdb3643acb0709b9d Oct 02 13:14:19 crc kubenswrapper[4710]: I1002 13:14:19.361705 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" event={"ID":"65b945dd-164b-45ec-9262-f5d7db0d15dc","Type":"ContainerStarted","Data":"e940c6b2782fc788f8ddc67cf00f2921c1a6b3b945e2517bdb3643acb0709b9d"} Oct 02 13:14:20 crc kubenswrapper[4710]: I1002 13:14:20.378155 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5bqq" event={"ID":"87b2eeb0-fdbf-4acb-9206-60e161ad6646","Type":"ContainerStarted","Data":"d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b"} Oct 02 13:14:20 crc kubenswrapper[4710]: I1002 13:14:20.394606 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d5bqq" podStartSLOduration=2.182844893 podStartE2EDuration="5.394584959s" podCreationTimestamp="2025-10-02 13:14:15 +0000 UTC" firstStartedPulling="2025-10-02 13:14:16.322806772 +0000 UTC m=+1500.429217655" lastFinishedPulling="2025-10-02 13:14:19.534546838 +0000 UTC m=+1503.640957721" observedRunningTime="2025-10-02 13:14:20.392566709 +0000 UTC m=+1504.498977602" watchObservedRunningTime="2025-10-02 13:14:20.394584959 +0000 UTC m=+1504.500995842" Oct 02 13:14:22 crc kubenswrapper[4710]: I1002 13:14:22.402651 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" event={"ID":"65b945dd-164b-45ec-9262-f5d7db0d15dc","Type":"ContainerStarted","Data":"a14ac22d4727a53bd0977868ca6ba57e3020de9f57f5c7c9a6f9299b1363a4e9"} Oct 02 13:14:25 crc kubenswrapper[4710]: I1002 13:14:25.384448 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:25 crc kubenswrapper[4710]: I1002 13:14:25.384829 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:25 crc kubenswrapper[4710]: I1002 13:14:25.431190 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:25 crc kubenswrapper[4710]: I1002 13:14:25.470928 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" podStartSLOduration=5.553644071 podStartE2EDuration="7.470903963s" podCreationTimestamp="2025-10-02 13:14:18 +0000 UTC" firstStartedPulling="2025-10-02 13:14:19.532948699 +0000 UTC m=+1503.639359592" lastFinishedPulling="2025-10-02 13:14:21.450208591 +0000 UTC m=+1505.556619484" observedRunningTime="2025-10-02 13:14:22.426099725 +0000 UTC m=+1506.532510618" watchObservedRunningTime="2025-10-02 13:14:25.470903963 +0000 UTC m=+1509.577314846" Oct 02 13:14:25 crc kubenswrapper[4710]: I1002 13:14:25.500649 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:25 crc kubenswrapper[4710]: I1002 13:14:25.675837 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d5bqq"] Oct 02 13:14:27 crc kubenswrapper[4710]: I1002 13:14:27.466432 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d5bqq" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="registry-server" containerID="cri-o://d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b" gracePeriod=2 Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.478274 4710 generic.go:334] "Generic (PLEG): container finished" podID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerID="d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b" exitCode=0 Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.478368 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.478361 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5bqq" event={"ID":"87b2eeb0-fdbf-4acb-9206-60e161ad6646","Type":"ContainerDied","Data":"d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b"} Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.478703 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5bqq" event={"ID":"87b2eeb0-fdbf-4acb-9206-60e161ad6646","Type":"ContainerDied","Data":"5a65c03ee42e0a12e57477e7e63069253071f53d99b57798f8ba38fae2b053cf"} Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.478728 4710 scope.go:117] "RemoveContainer" containerID="d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.511220 4710 scope.go:117] "RemoveContainer" containerID="87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.534701 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-utilities\") pod \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.534885 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrc6c\" (UniqueName: \"kubernetes.io/projected/87b2eeb0-fdbf-4acb-9206-60e161ad6646-kube-api-access-nrc6c\") pod \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.535162 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-catalog-content\") pod \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\" (UID: \"87b2eeb0-fdbf-4acb-9206-60e161ad6646\") " Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.536031 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-utilities" (OuterVolumeSpecName: "utilities") pod "87b2eeb0-fdbf-4acb-9206-60e161ad6646" (UID: "87b2eeb0-fdbf-4acb-9206-60e161ad6646"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.537852 4710 scope.go:117] "RemoveContainer" containerID="4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.543784 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b2eeb0-fdbf-4acb-9206-60e161ad6646-kube-api-access-nrc6c" (OuterVolumeSpecName: "kube-api-access-nrc6c") pod "87b2eeb0-fdbf-4acb-9206-60e161ad6646" (UID: "87b2eeb0-fdbf-4acb-9206-60e161ad6646"). InnerVolumeSpecName "kube-api-access-nrc6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.585588 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87b2eeb0-fdbf-4acb-9206-60e161ad6646" (UID: "87b2eeb0-fdbf-4acb-9206-60e161ad6646"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.637875 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.637920 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2eeb0-fdbf-4acb-9206-60e161ad6646-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.637931 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrc6c\" (UniqueName: \"kubernetes.io/projected/87b2eeb0-fdbf-4acb-9206-60e161ad6646-kube-api-access-nrc6c\") on node \"crc\" DevicePath \"\"" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.640529 4710 scope.go:117] "RemoveContainer" containerID="d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b" Oct 02 13:14:28 crc kubenswrapper[4710]: E1002 13:14:28.641161 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b\": container with ID starting with d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b not found: ID does not exist" containerID="d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.641204 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b"} err="failed to get container status \"d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b\": rpc error: code = NotFound desc = could not find container \"d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b\": container with ID starting with d0accdd49f986fbf0f0717b6329132f9f3c92c79f7106f09f23530b10321d86b not found: ID does not exist" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.641228 4710 scope.go:117] "RemoveContainer" containerID="87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd" Oct 02 13:14:28 crc kubenswrapper[4710]: E1002 13:14:28.642670 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd\": container with ID starting with 87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd not found: ID does not exist" containerID="87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.642709 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd"} err="failed to get container status \"87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd\": rpc error: code = NotFound desc = could not find container \"87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd\": container with ID starting with 87f9594aff73d666a31bca1cdddd1c9cd99502539487ef8f07d4fd8c336fd2cd not found: ID does not exist" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.642737 4710 scope.go:117] "RemoveContainer" containerID="4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9" Oct 02 13:14:28 crc kubenswrapper[4710]: E1002 13:14:28.643052 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9\": container with ID starting with 4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9 not found: ID does not exist" containerID="4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9" Oct 02 13:14:28 crc kubenswrapper[4710]: I1002 13:14:28.643078 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9"} err="failed to get container status \"4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9\": rpc error: code = NotFound desc = could not find container \"4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9\": container with ID starting with 4cd3db1d26a53b9572eae9948dad0c5ae21732fe84152de372ba1ba55c27a6a9 not found: ID does not exist" Oct 02 13:14:29 crc kubenswrapper[4710]: I1002 13:14:29.493296 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5bqq" Oct 02 13:14:29 crc kubenswrapper[4710]: I1002 13:14:29.518820 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d5bqq"] Oct 02 13:14:29 crc kubenswrapper[4710]: I1002 13:14:29.528044 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d5bqq"] Oct 02 13:14:30 crc kubenswrapper[4710]: I1002 13:14:30.918787 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" path="/var/lib/kubelet/pods/87b2eeb0-fdbf-4acb-9206-60e161ad6646/volumes" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.249580 4710 scope.go:117] "RemoveContainer" containerID="d599285d2de11783bf93fa6dc9c6be10e9a0fe21db62e996d8ec62843903b065" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.282214 4710 scope.go:117] "RemoveContainer" containerID="a42ca0c2045d5f637fdd39fdecb14ac1b0dd3e6fe65cc97a6a353834d22c6f67" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.307754 4710 scope.go:117] "RemoveContainer" containerID="b5c52f2790e18500ac8222c04e234a2e4b874932f58a260273793e7b95afe6ef" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.329899 4710 scope.go:117] "RemoveContainer" containerID="1dc130c3972d51c15bd9ed2746f526598c726a4f4dcc16e5440157a94ad252e1" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.393613 4710 scope.go:117] "RemoveContainer" containerID="2b4a4bb8893f083849330b9ab1404fe925487d732359bb18512233446d936a8b" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.572473 4710 scope.go:117] "RemoveContainer" containerID="f31d2c3b09a941fffe2619fe6a392f8621451418117fb3f7673b22acebccc053" Oct 02 13:14:44 crc kubenswrapper[4710]: I1002 13:14:44.599224 4710 scope.go:117] "RemoveContainer" containerID="ec2e13a9bfec1edcca0d4a80db652e7b9929f4e2b1328f686fedcb3856ec00a3" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.143494 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56"] Oct 02 13:15:00 crc kubenswrapper[4710]: E1002 13:15:00.144603 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="registry-server" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.144623 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="registry-server" Oct 02 13:15:00 crc kubenswrapper[4710]: E1002 13:15:00.144646 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="extract-utilities" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.144655 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="extract-utilities" Oct 02 13:15:00 crc kubenswrapper[4710]: E1002 13:15:00.144703 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="extract-content" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.144711 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="extract-content" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.144970 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b2eeb0-fdbf-4acb-9206-60e161ad6646" containerName="registry-server" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.145813 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.147860 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.148502 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.154330 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56"] Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.156969 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5569de9b-ca9d-4d55-a762-268ba7881c0b-config-volume\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.157010 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5569de9b-ca9d-4d55-a762-268ba7881c0b-secret-volume\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.157134 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62zhm\" (UniqueName: \"kubernetes.io/projected/5569de9b-ca9d-4d55-a762-268ba7881c0b-kube-api-access-62zhm\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.258321 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62zhm\" (UniqueName: \"kubernetes.io/projected/5569de9b-ca9d-4d55-a762-268ba7881c0b-kube-api-access-62zhm\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.258424 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5569de9b-ca9d-4d55-a762-268ba7881c0b-secret-volume\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.258450 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5569de9b-ca9d-4d55-a762-268ba7881c0b-config-volume\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.259629 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5569de9b-ca9d-4d55-a762-268ba7881c0b-config-volume\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.265487 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5569de9b-ca9d-4d55-a762-268ba7881c0b-secret-volume\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.280552 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62zhm\" (UniqueName: \"kubernetes.io/projected/5569de9b-ca9d-4d55-a762-268ba7881c0b-kube-api-access-62zhm\") pod \"collect-profiles-29323515-n2t56\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.471125 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:00 crc kubenswrapper[4710]: I1002 13:15:00.937180 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56"] Oct 02 13:15:01 crc kubenswrapper[4710]: I1002 13:15:01.824041 4710 generic.go:334] "Generic (PLEG): container finished" podID="5569de9b-ca9d-4d55-a762-268ba7881c0b" containerID="162c8a09c0856d03a3499f0d3854d21da87d4924247b41cb13568aba073aed58" exitCode=0 Oct 02 13:15:01 crc kubenswrapper[4710]: I1002 13:15:01.824091 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" event={"ID":"5569de9b-ca9d-4d55-a762-268ba7881c0b","Type":"ContainerDied","Data":"162c8a09c0856d03a3499f0d3854d21da87d4924247b41cb13568aba073aed58"} Oct 02 13:15:01 crc kubenswrapper[4710]: I1002 13:15:01.824362 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" event={"ID":"5569de9b-ca9d-4d55-a762-268ba7881c0b","Type":"ContainerStarted","Data":"fcf2786bfcd79a8889e8ee353dfab2a2864ad8a4b6d640a9c24ff308adc718ac"} Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.181526 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.212440 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5569de9b-ca9d-4d55-a762-268ba7881c0b-config-volume\") pod \"5569de9b-ca9d-4d55-a762-268ba7881c0b\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.212506 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62zhm\" (UniqueName: \"kubernetes.io/projected/5569de9b-ca9d-4d55-a762-268ba7881c0b-kube-api-access-62zhm\") pod \"5569de9b-ca9d-4d55-a762-268ba7881c0b\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.212624 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5569de9b-ca9d-4d55-a762-268ba7881c0b-secret-volume\") pod \"5569de9b-ca9d-4d55-a762-268ba7881c0b\" (UID: \"5569de9b-ca9d-4d55-a762-268ba7881c0b\") " Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.219068 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5569de9b-ca9d-4d55-a762-268ba7881c0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "5569de9b-ca9d-4d55-a762-268ba7881c0b" (UID: "5569de9b-ca9d-4d55-a762-268ba7881c0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.219204 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5569de9b-ca9d-4d55-a762-268ba7881c0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5569de9b-ca9d-4d55-a762-268ba7881c0b" (UID: "5569de9b-ca9d-4d55-a762-268ba7881c0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.221372 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5569de9b-ca9d-4d55-a762-268ba7881c0b-kube-api-access-62zhm" (OuterVolumeSpecName: "kube-api-access-62zhm") pod "5569de9b-ca9d-4d55-a762-268ba7881c0b" (UID: "5569de9b-ca9d-4d55-a762-268ba7881c0b"). InnerVolumeSpecName "kube-api-access-62zhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.314641 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5569de9b-ca9d-4d55-a762-268ba7881c0b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.315179 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62zhm\" (UniqueName: \"kubernetes.io/projected/5569de9b-ca9d-4d55-a762-268ba7881c0b-kube-api-access-62zhm\") on node \"crc\" DevicePath \"\"" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.315495 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5569de9b-ca9d-4d55-a762-268ba7881c0b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.845341 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" event={"ID":"5569de9b-ca9d-4d55-a762-268ba7881c0b","Type":"ContainerDied","Data":"fcf2786bfcd79a8889e8ee353dfab2a2864ad8a4b6d640a9c24ff308adc718ac"} Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.845389 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcf2786bfcd79a8889e8ee353dfab2a2864ad8a4b6d640a9c24ff308adc718ac" Oct 02 13:15:03 crc kubenswrapper[4710]: I1002 13:15:03.845686 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56" Oct 02 13:15:22 crc kubenswrapper[4710]: I1002 13:15:22.530155 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:15:22 crc kubenswrapper[4710]: I1002 13:15:22.530760 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:15:52 crc kubenswrapper[4710]: I1002 13:15:52.530050 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:15:52 crc kubenswrapper[4710]: I1002 13:15:52.530580 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:16:11 crc kubenswrapper[4710]: I1002 13:16:11.062257 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-hh65j"] Oct 02 13:16:11 crc kubenswrapper[4710]: I1002 13:16:11.080953 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-hh65j"] Oct 02 13:16:12 crc kubenswrapper[4710]: I1002 13:16:12.921112 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdc93b0-6465-43e9-8d1f-045f2ee24460" path="/var/lib/kubelet/pods/afdc93b0-6465-43e9-8d1f-045f2ee24460/volumes" Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.031480 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-hzh54"] Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.041077 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-hzh54"] Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.049870 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-79f7g"] Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.057958 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-msp99"] Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.065605 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-msp99"] Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.073117 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-79f7g"] Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.915373 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53530783-8e7c-4e44-9484-6e5e9ae70bd2" path="/var/lib/kubelet/pods/53530783-8e7c-4e44-9484-6e5e9ae70bd2/volumes" Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.916248 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ed1b30c-53d5-437e-b5c2-d5526668c15d" path="/var/lib/kubelet/pods/5ed1b30c-53d5-437e-b5c2-d5526668c15d/volumes" Oct 02 13:16:14 crc kubenswrapper[4710]: I1002 13:16:14.916865 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdea4fdc-8bd8-4a6a-afa1-e0574adfce51" path="/var/lib/kubelet/pods/fdea4fdc-8bd8-4a6a-afa1-e0574adfce51/volumes" Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.027186 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c3fa-account-create-ml5qg"] Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.035893 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c3fa-account-create-ml5qg"] Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.530203 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.530278 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.530333 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.531559 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.531635 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" gracePeriod=600 Oct 02 13:16:22 crc kubenswrapper[4710]: E1002 13:16:22.792434 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:16:22 crc kubenswrapper[4710]: I1002 13:16:22.915521 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f96697-8eea-4ade-a5f4-a51986a0f83f" path="/var/lib/kubelet/pods/d1f96697-8eea-4ade-a5f4-a51986a0f83f/volumes" Oct 02 13:16:23 crc kubenswrapper[4710]: I1002 13:16:23.663986 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" exitCode=0 Oct 02 13:16:23 crc kubenswrapper[4710]: I1002 13:16:23.664038 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e"} Oct 02 13:16:23 crc kubenswrapper[4710]: I1002 13:16:23.664075 4710 scope.go:117] "RemoveContainer" containerID="4b5183e50d1f99d425b43adb71d4ff8d51ea84c4eb7c8acc094921b9bbeee6bd" Oct 02 13:16:23 crc kubenswrapper[4710]: I1002 13:16:23.664944 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:16:23 crc kubenswrapper[4710]: E1002 13:16:23.665314 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:16:24 crc kubenswrapper[4710]: I1002 13:16:24.031688 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-7e6f-account-create-8tfx4"] Oct 02 13:16:24 crc kubenswrapper[4710]: I1002 13:16:24.043045 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-7e6f-account-create-8tfx4"] Oct 02 13:16:24 crc kubenswrapper[4710]: I1002 13:16:24.918100 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="882558ba-032b-4377-a806-ae3760562289" path="/var/lib/kubelet/pods/882558ba-032b-4377-a806-ae3760562289/volumes" Oct 02 13:16:27 crc kubenswrapper[4710]: I1002 13:16:27.030250 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fdf1-account-create-9z6r2"] Oct 02 13:16:27 crc kubenswrapper[4710]: I1002 13:16:27.038815 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fdf1-account-create-9z6r2"] Oct 02 13:16:28 crc kubenswrapper[4710]: I1002 13:16:28.919208 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd700f13-8732-4f7f-bbaf-ba5f8156ba97" path="/var/lib/kubelet/pods/dd700f13-8732-4f7f-bbaf-ba5f8156ba97/volumes" Oct 02 13:16:32 crc kubenswrapper[4710]: I1002 13:16:32.045093 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7d43-account-create-d6l9p"] Oct 02 13:16:32 crc kubenswrapper[4710]: I1002 13:16:32.054003 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7d43-account-create-d6l9p"] Oct 02 13:16:32 crc kubenswrapper[4710]: I1002 13:16:32.918062 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce991bb5-1e98-40f5-8395-b1997433b236" path="/var/lib/kubelet/pods/ce991bb5-1e98-40f5-8395-b1997433b236/volumes" Oct 02 13:16:36 crc kubenswrapper[4710]: I1002 13:16:36.911352 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:16:36 crc kubenswrapper[4710]: E1002 13:16:36.912221 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:16:44 crc kubenswrapper[4710]: I1002 13:16:44.784709 4710 scope.go:117] "RemoveContainer" containerID="dafafd991dee79759d06a6a546daf02dce167312d311979e539fed3a6e71c841" Oct 02 13:16:44 crc kubenswrapper[4710]: I1002 13:16:44.813314 4710 scope.go:117] "RemoveContainer" containerID="85a23d9e67017a47c214124a314e55456698490a9ce07b92aa893470b2979bbb" Oct 02 13:16:44 crc kubenswrapper[4710]: I1002 13:16:44.866104 4710 scope.go:117] "RemoveContainer" containerID="4712bbc24352470c7207735dbfbd436fc4b312e6d98f93bf18c31b64577d6747" Oct 02 13:16:44 crc kubenswrapper[4710]: I1002 13:16:44.909925 4710 scope.go:117] "RemoveContainer" containerID="f228ccc4d1c2968d9fd2fc7f67215541470459872262a71452cec41c7c69d6ad" Oct 02 13:16:44 crc kubenswrapper[4710]: I1002 13:16:44.954856 4710 scope.go:117] "RemoveContainer" containerID="efa390cf6d953d43af790b8552702466050dce517b3cee5e250e1061e5c4c296" Oct 02 13:16:45 crc kubenswrapper[4710]: I1002 13:16:45.011803 4710 scope.go:117] "RemoveContainer" containerID="19f69cc552f88b7ed533dc2618f92ba1e0d58eb7ae0f5fd9330bbc19a0738005" Oct 02 13:16:45 crc kubenswrapper[4710]: I1002 13:16:45.078234 4710 scope.go:117] "RemoveContainer" containerID="b48cb2a23135d3dd9b627dd5b80630befe8bf029fdd2fc9a170428e47600bc25" Oct 02 13:16:45 crc kubenswrapper[4710]: I1002 13:16:45.104064 4710 scope.go:117] "RemoveContainer" containerID="a793ca2ff867931392f02910af50c697e978cc00a899303389ea43e4c72f83df" Oct 02 13:16:48 crc kubenswrapper[4710]: I1002 13:16:48.904915 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:16:48 crc kubenswrapper[4710]: E1002 13:16:48.905783 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:16:51 crc kubenswrapper[4710]: I1002 13:16:51.038922 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-b4nsw"] Oct 02 13:16:51 crc kubenswrapper[4710]: I1002 13:16:51.050254 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-b4nsw"] Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.046795 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-wc99n"] Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.047850 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-cw6sh"] Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.060852 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-wc99n"] Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.072124 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-cw6sh"] Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.928061 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25de8c63-18f4-4896-b856-3c5c20d2d54a" path="/var/lib/kubelet/pods/25de8c63-18f4-4896-b856-3c5c20d2d54a/volumes" Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.929149 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe" path="/var/lib/kubelet/pods/2fa748a4-55a5-4f2b-99f9-6d049d7c4bbe/volumes" Oct 02 13:16:52 crc kubenswrapper[4710]: I1002 13:16:52.929657 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a4c671-5a25-4cd5-9b19-db54e931b28b" path="/var/lib/kubelet/pods/79a4c671-5a25-4cd5-9b19-db54e931b28b/volumes" Oct 02 13:17:01 crc kubenswrapper[4710]: I1002 13:17:01.031352 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-6htbd"] Oct 02 13:17:01 crc kubenswrapper[4710]: I1002 13:17:01.038368 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-6htbd"] Oct 02 13:17:02 crc kubenswrapper[4710]: I1002 13:17:02.918858 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52ce5f44-dff4-4fe5-aa15-e76e352b9d88" path="/var/lib/kubelet/pods/52ce5f44-dff4-4fe5-aa15-e76e352b9d88/volumes" Oct 02 13:17:03 crc kubenswrapper[4710]: I1002 13:17:03.905206 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:17:03 crc kubenswrapper[4710]: E1002 13:17:03.905486 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:17:05 crc kubenswrapper[4710]: I1002 13:17:05.033770 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9aae-account-create-vt2hh"] Oct 02 13:17:05 crc kubenswrapper[4710]: I1002 13:17:05.048414 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-21d4-account-create-r2d77"] Oct 02 13:17:05 crc kubenswrapper[4710]: I1002 13:17:05.060325 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-a0e6-account-create-n8wcs"] Oct 02 13:17:05 crc kubenswrapper[4710]: I1002 13:17:05.078943 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-21d4-account-create-r2d77"] Oct 02 13:17:05 crc kubenswrapper[4710]: I1002 13:17:05.088125 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9aae-account-create-vt2hh"] Oct 02 13:17:05 crc kubenswrapper[4710]: I1002 13:17:05.097377 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-a0e6-account-create-n8wcs"] Oct 02 13:17:06 crc kubenswrapper[4710]: I1002 13:17:06.916483 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03930039-3e8f-4a37-ae83-61f08a70aea4" path="/var/lib/kubelet/pods/03930039-3e8f-4a37-ae83-61f08a70aea4/volumes" Oct 02 13:17:06 crc kubenswrapper[4710]: I1002 13:17:06.917363 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34948622-b486-47f0-bbb7-832f5ca6a1e7" path="/var/lib/kubelet/pods/34948622-b486-47f0-bbb7-832f5ca6a1e7/volumes" Oct 02 13:17:06 crc kubenswrapper[4710]: I1002 13:17:06.918001 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39dc6a17-1fa2-440c-9ee4-3b12755fc422" path="/var/lib/kubelet/pods/39dc6a17-1fa2-440c-9ee4-3b12755fc422/volumes" Oct 02 13:17:09 crc kubenswrapper[4710]: I1002 13:17:09.040122 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-cm5dq"] Oct 02 13:17:09 crc kubenswrapper[4710]: I1002 13:17:09.049309 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-cm5dq"] Oct 02 13:17:10 crc kubenswrapper[4710]: I1002 13:17:10.036154 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xf9wb"] Oct 02 13:17:10 crc kubenswrapper[4710]: I1002 13:17:10.044553 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xf9wb"] Oct 02 13:17:10 crc kubenswrapper[4710]: I1002 13:17:10.914726 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="765efee7-b0c7-417f-8ad6-fee3d089bc50" path="/var/lib/kubelet/pods/765efee7-b0c7-417f-8ad6-fee3d089bc50/volumes" Oct 02 13:17:10 crc kubenswrapper[4710]: I1002 13:17:10.915461 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad3b2f5f-f4d4-4a80-948d-275033ab9f05" path="/var/lib/kubelet/pods/ad3b2f5f-f4d4-4a80-948d-275033ab9f05/volumes" Oct 02 13:17:14 crc kubenswrapper[4710]: I1002 13:17:14.908682 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:17:14 crc kubenswrapper[4710]: E1002 13:17:14.909081 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:17:28 crc kubenswrapper[4710]: I1002 13:17:28.904540 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:17:28 crc kubenswrapper[4710]: E1002 13:17:28.905228 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:17:39 crc kubenswrapper[4710]: I1002 13:17:39.904770 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:17:39 crc kubenswrapper[4710]: E1002 13:17:39.905507 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.262261 4710 scope.go:117] "RemoveContainer" containerID="a36bbda7761dac810ffb248a657d9ec4e6e6e94fd02510553c6b3d9e12c84cde" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.299192 4710 scope.go:117] "RemoveContainer" containerID="9fdf852541c55aa2bc18ed6b6bedc735d9f85d14c120c89b2d22fbf99c473b63" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.372688 4710 scope.go:117] "RemoveContainer" containerID="b717f43396ad4516ea52f0f94c6eeab96522ec9934f263714841fd6374972f31" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.440195 4710 scope.go:117] "RemoveContainer" containerID="4a41bdc2ba6285c7733c4b96675ea15361dc0b06aaf8cb2059defb06e978fa4d" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.477178 4710 scope.go:117] "RemoveContainer" containerID="23d65eded6f15a0f22f1963e12678f7c593c49010d2f67512979d07691a537a6" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.554686 4710 scope.go:117] "RemoveContainer" containerID="2b30bf42c58fda06f2f6f5367a2b184adbcbbe45eb40d08292630d7659fef79e" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.590283 4710 scope.go:117] "RemoveContainer" containerID="9b3950de6e8766c1c2d5fe1bb14dfc026c167355de3b5946ddf9c8e01add197f" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.612739 4710 scope.go:117] "RemoveContainer" containerID="78a723072bd56a19fb6d40a5635d3457b5b60090db8f1c0692d0cd2d73db4adf" Oct 02 13:17:45 crc kubenswrapper[4710]: I1002 13:17:45.655589 4710 scope.go:117] "RemoveContainer" containerID="2a3b4974ea13ea15af50dfac12ed9bf8640b4b508b0ab89e17d5f03c364b5a17" Oct 02 13:17:47 crc kubenswrapper[4710]: I1002 13:17:47.478429 4710 generic.go:334] "Generic (PLEG): container finished" podID="65b945dd-164b-45ec-9262-f5d7db0d15dc" containerID="a14ac22d4727a53bd0977868ca6ba57e3020de9f57f5c7c9a6f9299b1363a4e9" exitCode=0 Oct 02 13:17:47 crc kubenswrapper[4710]: I1002 13:17:47.478398 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" event={"ID":"65b945dd-164b-45ec-9262-f5d7db0d15dc","Type":"ContainerDied","Data":"a14ac22d4727a53bd0977868ca6ba57e3020de9f57f5c7c9a6f9299b1363a4e9"} Oct 02 13:17:47 crc kubenswrapper[4710]: E1002 13:17:47.507694 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65b945dd_164b_45ec_9262_f5d7db0d15dc.slice/crio-a14ac22d4727a53bd0977868ca6ba57e3020de9f57f5c7c9a6f9299b1363a4e9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65b945dd_164b_45ec_9262_f5d7db0d15dc.slice/crio-conmon-a14ac22d4727a53bd0977868ca6ba57e3020de9f57f5c7c9a6f9299b1363a4e9.scope\": RecentStats: unable to find data in memory cache]" Oct 02 13:17:48 crc kubenswrapper[4710]: I1002 13:17:48.986109 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.057485 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxxj6\" (UniqueName: \"kubernetes.io/projected/65b945dd-164b-45ec-9262-f5d7db0d15dc-kube-api-access-dxxj6\") pod \"65b945dd-164b-45ec-9262-f5d7db0d15dc\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.057585 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-inventory\") pod \"65b945dd-164b-45ec-9262-f5d7db0d15dc\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.058300 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-ssh-key\") pod \"65b945dd-164b-45ec-9262-f5d7db0d15dc\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.058470 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-bootstrap-combined-ca-bundle\") pod \"65b945dd-164b-45ec-9262-f5d7db0d15dc\" (UID: \"65b945dd-164b-45ec-9262-f5d7db0d15dc\") " Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.064074 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b945dd-164b-45ec-9262-f5d7db0d15dc-kube-api-access-dxxj6" (OuterVolumeSpecName: "kube-api-access-dxxj6") pod "65b945dd-164b-45ec-9262-f5d7db0d15dc" (UID: "65b945dd-164b-45ec-9262-f5d7db0d15dc"). InnerVolumeSpecName "kube-api-access-dxxj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.064218 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "65b945dd-164b-45ec-9262-f5d7db0d15dc" (UID: "65b945dd-164b-45ec-9262-f5d7db0d15dc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.086091 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "65b945dd-164b-45ec-9262-f5d7db0d15dc" (UID: "65b945dd-164b-45ec-9262-f5d7db0d15dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.089808 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-inventory" (OuterVolumeSpecName: "inventory") pod "65b945dd-164b-45ec-9262-f5d7db0d15dc" (UID: "65b945dd-164b-45ec-9262-f5d7db0d15dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.160362 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.160396 4710 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.160440 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxxj6\" (UniqueName: \"kubernetes.io/projected/65b945dd-164b-45ec-9262-f5d7db0d15dc-kube-api-access-dxxj6\") on node \"crc\" DevicePath \"\"" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.160450 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65b945dd-164b-45ec-9262-f5d7db0d15dc-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.501186 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" event={"ID":"65b945dd-164b-45ec-9262-f5d7db0d15dc","Type":"ContainerDied","Data":"e940c6b2782fc788f8ddc67cf00f2921c1a6b3b945e2517bdb3643acb0709b9d"} Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.501215 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.501227 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e940c6b2782fc788f8ddc67cf00f2921c1a6b3b945e2517bdb3643acb0709b9d" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.582269 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7"] Oct 02 13:17:49 crc kubenswrapper[4710]: E1002 13:17:49.582732 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b945dd-164b-45ec-9262-f5d7db0d15dc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.582996 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b945dd-164b-45ec-9262-f5d7db0d15dc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 13:17:49 crc kubenswrapper[4710]: E1002 13:17:49.583058 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5569de9b-ca9d-4d55-a762-268ba7881c0b" containerName="collect-profiles" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.583067 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="5569de9b-ca9d-4d55-a762-268ba7881c0b" containerName="collect-profiles" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.583339 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="5569de9b-ca9d-4d55-a762-268ba7881c0b" containerName="collect-profiles" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.583362 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b945dd-164b-45ec-9262-f5d7db0d15dc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.584231 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.587139 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.587288 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.587576 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.592522 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7"] Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.594526 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.672327 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.672787 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnsl4\" (UniqueName: \"kubernetes.io/projected/74b5366f-d978-41b8-85e6-a9bdb91230ce-kube-api-access-cnsl4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.673153 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.775707 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.775922 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.776133 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnsl4\" (UniqueName: \"kubernetes.io/projected/74b5366f-d978-41b8-85e6-a9bdb91230ce-kube-api-access-cnsl4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.779819 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.779912 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.792902 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnsl4\" (UniqueName: \"kubernetes.io/projected/74b5366f-d978-41b8-85e6-a9bdb91230ce-kube-api-access-cnsl4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:49 crc kubenswrapper[4710]: I1002 13:17:49.909036 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:17:50 crc kubenswrapper[4710]: I1002 13:17:50.449514 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7"] Oct 02 13:17:50 crc kubenswrapper[4710]: I1002 13:17:50.511384 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" event={"ID":"74b5366f-d978-41b8-85e6-a9bdb91230ce","Type":"ContainerStarted","Data":"40db99a10033c85f88ecd12a3735551ea0cd91d6af22f9c2abb8e0cc86c87e53"} Oct 02 13:17:52 crc kubenswrapper[4710]: I1002 13:17:52.529505 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" event={"ID":"74b5366f-d978-41b8-85e6-a9bdb91230ce","Type":"ContainerStarted","Data":"5ddf40f29cdfeffb918c781659bd1fcf23be5cd2fcfcbc75a5af86e35d7a585e"} Oct 02 13:17:52 crc kubenswrapper[4710]: I1002 13:17:52.543797 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" podStartSLOduration=2.567445965 podStartE2EDuration="3.543777576s" podCreationTimestamp="2025-10-02 13:17:49 +0000 UTC" firstStartedPulling="2025-10-02 13:17:50.460248785 +0000 UTC m=+1714.566659688" lastFinishedPulling="2025-10-02 13:17:51.436580416 +0000 UTC m=+1715.542991299" observedRunningTime="2025-10-02 13:17:52.542583857 +0000 UTC m=+1716.648994740" watchObservedRunningTime="2025-10-02 13:17:52.543777576 +0000 UTC m=+1716.650188459" Oct 02 13:17:53 crc kubenswrapper[4710]: I1002 13:17:53.904720 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:17:53 crc kubenswrapper[4710]: E1002 13:17:53.906610 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:17:55 crc kubenswrapper[4710]: I1002 13:17:55.047972 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-987x9"] Oct 02 13:17:55 crc kubenswrapper[4710]: I1002 13:17:55.059040 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dqkk2"] Oct 02 13:17:55 crc kubenswrapper[4710]: I1002 13:17:55.071191 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dqkk2"] Oct 02 13:17:55 crc kubenswrapper[4710]: I1002 13:17:55.081621 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-987x9"] Oct 02 13:17:56 crc kubenswrapper[4710]: I1002 13:17:56.919227 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54924b84-6352-4ea1-8e55-35ebc23fc0a0" path="/var/lib/kubelet/pods/54924b84-6352-4ea1-8e55-35ebc23fc0a0/volumes" Oct 02 13:17:56 crc kubenswrapper[4710]: I1002 13:17:56.921061 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6216b148-2847-437c-8437-1f574973fb87" path="/var/lib/kubelet/pods/6216b148-2847-437c-8437-1f574973fb87/volumes" Oct 02 13:17:58 crc kubenswrapper[4710]: I1002 13:17:58.039737 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-swkp8"] Oct 02 13:17:58 crc kubenswrapper[4710]: I1002 13:17:58.052031 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-swkp8"] Oct 02 13:17:58 crc kubenswrapper[4710]: I1002 13:17:58.918687 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978ba8dd-e49f-40cc-a4f9-4dd5658e4001" path="/var/lib/kubelet/pods/978ba8dd-e49f-40cc-a4f9-4dd5658e4001/volumes" Oct 02 13:18:04 crc kubenswrapper[4710]: I1002 13:18:04.904936 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:18:04 crc kubenswrapper[4710]: E1002 13:18:04.906082 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:18:18 crc kubenswrapper[4710]: I1002 13:18:18.904480 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:18:18 crc kubenswrapper[4710]: E1002 13:18:18.905403 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:18:21 crc kubenswrapper[4710]: I1002 13:18:21.039887 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-nwcvq"] Oct 02 13:18:21 crc kubenswrapper[4710]: I1002 13:18:21.049084 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-nwcvq"] Oct 02 13:18:22 crc kubenswrapper[4710]: I1002 13:18:22.916629 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f53ffe3d-d27a-41dd-a98b-4debca318ec4" path="/var/lib/kubelet/pods/f53ffe3d-d27a-41dd-a98b-4debca318ec4/volumes" Oct 02 13:18:26 crc kubenswrapper[4710]: I1002 13:18:26.033877 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4m9cc"] Oct 02 13:18:26 crc kubenswrapper[4710]: I1002 13:18:26.044904 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4m9cc"] Oct 02 13:18:26 crc kubenswrapper[4710]: I1002 13:18:26.915889 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6152e5af-f395-4bca-9aa1-19d17ec0ffa2" path="/var/lib/kubelet/pods/6152e5af-f395-4bca-9aa1-19d17ec0ffa2/volumes" Oct 02 13:18:32 crc kubenswrapper[4710]: I1002 13:18:32.904810 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:18:32 crc kubenswrapper[4710]: E1002 13:18:32.905788 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:18:45 crc kubenswrapper[4710]: I1002 13:18:45.882209 4710 scope.go:117] "RemoveContainer" containerID="6cb0c0e41cbdc2b0ef9923b55735f52424c85b6978f3aeb14939de427c0c5873" Oct 02 13:18:45 crc kubenswrapper[4710]: I1002 13:18:45.909967 4710 scope.go:117] "RemoveContainer" containerID="85cae4838ac68ae9e6883f3dc5e4fe695b34b8fe3135f482a2d580ac25cec23c" Oct 02 13:18:45 crc kubenswrapper[4710]: I1002 13:18:45.938663 4710 scope.go:117] "RemoveContainer" containerID="712b61b517d05149b02904aa2054f336915e1e5e34db2dae2941851eac765724" Oct 02 13:18:46 crc kubenswrapper[4710]: I1002 13:18:46.012570 4710 scope.go:117] "RemoveContainer" containerID="22317eba05ad5d22f364de32e34802eb9dddea15d9c3ac8122e08d32a81638fc" Oct 02 13:18:46 crc kubenswrapper[4710]: I1002 13:18:46.069778 4710 scope.go:117] "RemoveContainer" containerID="0699a8b45c702e7fd6ac9e39accb03b239a91dbb8189fd8c854e6048a25d80d9" Oct 02 13:18:46 crc kubenswrapper[4710]: I1002 13:18:46.142254 4710 scope.go:117] "RemoveContainer" containerID="f542341d24d05eeef5796236558be49881216b55d324ae55847b386a0ae3b051" Oct 02 13:18:46 crc kubenswrapper[4710]: I1002 13:18:46.174929 4710 scope.go:117] "RemoveContainer" containerID="c192fad0aa3ecc9de8b9978f169495111f49b8a3373adf52227e95e82935aea5" Oct 02 13:18:46 crc kubenswrapper[4710]: I1002 13:18:46.215205 4710 scope.go:117] "RemoveContainer" containerID="258fdf3876f9e30b2aeb09f882a559bc54129f5529eed0bc6e82425ed444a55a" Oct 02 13:18:46 crc kubenswrapper[4710]: I1002 13:18:46.911339 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:18:46 crc kubenswrapper[4710]: E1002 13:18:46.911651 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:18:57 crc kubenswrapper[4710]: I1002 13:18:57.904620 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:18:57 crc kubenswrapper[4710]: E1002 13:18:57.905511 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:19:09 crc kubenswrapper[4710]: I1002 13:19:09.915760 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:19:09 crc kubenswrapper[4710]: E1002 13:19:09.916668 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:19:21 crc kubenswrapper[4710]: I1002 13:19:21.904638 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:19:21 crc kubenswrapper[4710]: E1002 13:19:21.905321 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:19:29 crc kubenswrapper[4710]: I1002 13:19:29.449408 4710 generic.go:334] "Generic (PLEG): container finished" podID="74b5366f-d978-41b8-85e6-a9bdb91230ce" containerID="5ddf40f29cdfeffb918c781659bd1fcf23be5cd2fcfcbc75a5af86e35d7a585e" exitCode=0 Oct 02 13:19:29 crc kubenswrapper[4710]: I1002 13:19:29.449510 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" event={"ID":"74b5366f-d978-41b8-85e6-a9bdb91230ce","Type":"ContainerDied","Data":"5ddf40f29cdfeffb918c781659bd1fcf23be5cd2fcfcbc75a5af86e35d7a585e"} Oct 02 13:19:30 crc kubenswrapper[4710]: I1002 13:19:30.912657 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.021898 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnsl4\" (UniqueName: \"kubernetes.io/projected/74b5366f-d978-41b8-85e6-a9bdb91230ce-kube-api-access-cnsl4\") pod \"74b5366f-d978-41b8-85e6-a9bdb91230ce\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.022044 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-inventory\") pod \"74b5366f-d978-41b8-85e6-a9bdb91230ce\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.022094 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-ssh-key\") pod \"74b5366f-d978-41b8-85e6-a9bdb91230ce\" (UID: \"74b5366f-d978-41b8-85e6-a9bdb91230ce\") " Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.027526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b5366f-d978-41b8-85e6-a9bdb91230ce-kube-api-access-cnsl4" (OuterVolumeSpecName: "kube-api-access-cnsl4") pod "74b5366f-d978-41b8-85e6-a9bdb91230ce" (UID: "74b5366f-d978-41b8-85e6-a9bdb91230ce"). InnerVolumeSpecName "kube-api-access-cnsl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.049257 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74b5366f-d978-41b8-85e6-a9bdb91230ce" (UID: "74b5366f-d978-41b8-85e6-a9bdb91230ce"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.054551 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-inventory" (OuterVolumeSpecName: "inventory") pod "74b5366f-d978-41b8-85e6-a9bdb91230ce" (UID: "74b5366f-d978-41b8-85e6-a9bdb91230ce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.145331 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.145403 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b5366f-d978-41b8-85e6-a9bdb91230ce-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.145418 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnsl4\" (UniqueName: \"kubernetes.io/projected/74b5366f-d978-41b8-85e6-a9bdb91230ce-kube-api-access-cnsl4\") on node \"crc\" DevicePath \"\"" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.468557 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" event={"ID":"74b5366f-d978-41b8-85e6-a9bdb91230ce","Type":"ContainerDied","Data":"40db99a10033c85f88ecd12a3735551ea0cd91d6af22f9c2abb8e0cc86c87e53"} Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.468606 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.468616 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40db99a10033c85f88ecd12a3735551ea0cd91d6af22f9c2abb8e0cc86c87e53" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.566505 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n"] Oct 02 13:19:31 crc kubenswrapper[4710]: E1002 13:19:31.567313 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b5366f-d978-41b8-85e6-a9bdb91230ce" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.567333 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b5366f-d978-41b8-85e6-a9bdb91230ce" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.567519 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b5366f-d978-41b8-85e6-a9bdb91230ce" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.568309 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.571445 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.572387 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.574334 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.574363 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.582520 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n"] Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.656187 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.656297 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.656346 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxz8j\" (UniqueName: \"kubernetes.io/projected/ef9d9d20-85b8-46c0-9696-e05d43c16d39-kube-api-access-qxz8j\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.758387 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.758466 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxz8j\" (UniqueName: \"kubernetes.io/projected/ef9d9d20-85b8-46c0-9696-e05d43c16d39-kube-api-access-qxz8j\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.758661 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.764012 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.775473 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.778693 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxz8j\" (UniqueName: \"kubernetes.io/projected/ef9d9d20-85b8-46c0-9696-e05d43c16d39-kube-api-access-qxz8j\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nf64n\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:31 crc kubenswrapper[4710]: I1002 13:19:31.886834 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:19:32 crc kubenswrapper[4710]: I1002 13:19:32.650872 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n"] Oct 02 13:19:32 crc kubenswrapper[4710]: I1002 13:19:32.652678 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:19:33 crc kubenswrapper[4710]: I1002 13:19:33.487200 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" event={"ID":"ef9d9d20-85b8-46c0-9696-e05d43c16d39","Type":"ContainerStarted","Data":"7ba02f36a5defcdccabb7fc2f1e5c1ec17857fc787bba554617416cb643ae837"} Oct 02 13:19:33 crc kubenswrapper[4710]: I1002 13:19:33.487685 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" event={"ID":"ef9d9d20-85b8-46c0-9696-e05d43c16d39","Type":"ContainerStarted","Data":"66619cdee2ffc510334cd78b573a03425b5e9bb8346e8ac027d2d39a27caf15e"} Oct 02 13:19:33 crc kubenswrapper[4710]: I1002 13:19:33.513467 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" podStartSLOduration=2.007300403 podStartE2EDuration="2.51338849s" podCreationTimestamp="2025-10-02 13:19:31 +0000 UTC" firstStartedPulling="2025-10-02 13:19:32.652393716 +0000 UTC m=+1816.758804599" lastFinishedPulling="2025-10-02 13:19:33.158481783 +0000 UTC m=+1817.264892686" observedRunningTime="2025-10-02 13:19:33.504960987 +0000 UTC m=+1817.611371880" watchObservedRunningTime="2025-10-02 13:19:33.51338849 +0000 UTC m=+1817.619799373" Oct 02 13:19:34 crc kubenswrapper[4710]: I1002 13:19:34.904869 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:19:34 crc kubenswrapper[4710]: E1002 13:19:34.905248 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:19:45 crc kubenswrapper[4710]: I1002 13:19:45.904793 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:19:45 crc kubenswrapper[4710]: E1002 13:19:45.905670 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:19:49 crc kubenswrapper[4710]: I1002 13:19:49.042675 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xq4sp"] Oct 02 13:19:49 crc kubenswrapper[4710]: I1002 13:19:49.054826 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-gwdnf"] Oct 02 13:19:49 crc kubenswrapper[4710]: I1002 13:19:49.062856 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lmkkc"] Oct 02 13:19:49 crc kubenswrapper[4710]: I1002 13:19:49.071175 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lmkkc"] Oct 02 13:19:49 crc kubenswrapper[4710]: I1002 13:19:49.078803 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xq4sp"] Oct 02 13:19:49 crc kubenswrapper[4710]: I1002 13:19:49.086119 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-gwdnf"] Oct 02 13:19:50 crc kubenswrapper[4710]: I1002 13:19:50.920361 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33245146-2264-4fcf-a29b-a0272f6936e9" path="/var/lib/kubelet/pods/33245146-2264-4fcf-a29b-a0272f6936e9/volumes" Oct 02 13:19:50 crc kubenswrapper[4710]: I1002 13:19:50.921484 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8b28f29-2760-43a5-b8fc-af19ad1a3bcf" path="/var/lib/kubelet/pods/b8b28f29-2760-43a5-b8fc-af19ad1a3bcf/volumes" Oct 02 13:19:50 crc kubenswrapper[4710]: I1002 13:19:50.922226 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be6c2122-9be7-4414-9e44-8a25e8984d40" path="/var/lib/kubelet/pods/be6c2122-9be7-4414-9e44-8a25e8984d40/volumes" Oct 02 13:19:56 crc kubenswrapper[4710]: I1002 13:19:56.912145 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:19:56 crc kubenswrapper[4710]: E1002 13:19:56.913081 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:19:59 crc kubenswrapper[4710]: I1002 13:19:59.033229 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-94ae-account-create-5r4pn"] Oct 02 13:19:59 crc kubenswrapper[4710]: I1002 13:19:59.043770 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-eb66-account-create-bcf9b"] Oct 02 13:19:59 crc kubenswrapper[4710]: I1002 13:19:59.053142 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-d881-account-create-nvqlz"] Oct 02 13:19:59 crc kubenswrapper[4710]: I1002 13:19:59.063205 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-eb66-account-create-bcf9b"] Oct 02 13:19:59 crc kubenswrapper[4710]: I1002 13:19:59.072249 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-94ae-account-create-5r4pn"] Oct 02 13:19:59 crc kubenswrapper[4710]: I1002 13:19:59.081708 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-d881-account-create-nvqlz"] Oct 02 13:20:00 crc kubenswrapper[4710]: I1002 13:20:00.915078 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f9bab2f-dc35-45c4-a7bd-fae3a3576042" path="/var/lib/kubelet/pods/0f9bab2f-dc35-45c4-a7bd-fae3a3576042/volumes" Oct 02 13:20:00 crc kubenswrapper[4710]: I1002 13:20:00.915956 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c8a8393-59d4-4c27-9800-d96b2bc96b42" path="/var/lib/kubelet/pods/1c8a8393-59d4-4c27-9800-d96b2bc96b42/volumes" Oct 02 13:20:00 crc kubenswrapper[4710]: I1002 13:20:00.916422 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63bbbc57-c9bc-43fd-8bd9-15845ff6cd47" path="/var/lib/kubelet/pods/63bbbc57-c9bc-43fd-8bd9-15845ff6cd47/volumes" Oct 02 13:20:10 crc kubenswrapper[4710]: I1002 13:20:10.905120 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:20:10 crc kubenswrapper[4710]: E1002 13:20:10.905946 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:20:24 crc kubenswrapper[4710]: I1002 13:20:24.905368 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:20:24 crc kubenswrapper[4710]: E1002 13:20:24.906129 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:20:35 crc kubenswrapper[4710]: I1002 13:20:35.904601 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:20:35 crc kubenswrapper[4710]: E1002 13:20:35.905462 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:20:36 crc kubenswrapper[4710]: I1002 13:20:36.037434 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l6gmn"] Oct 02 13:20:36 crc kubenswrapper[4710]: I1002 13:20:36.065696 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l6gmn"] Oct 02 13:20:36 crc kubenswrapper[4710]: I1002 13:20:36.915408 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3004d0-bcfc-490a-bc66-0b14469fb0a5" path="/var/lib/kubelet/pods/ff3004d0-bcfc-490a-bc66-0b14469fb0a5/volumes" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.392449 4710 scope.go:117] "RemoveContainer" containerID="c2df07b5e6a8bff7df42122cbea490fef586921805eeb66fbdd30ea523845b96" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.415446 4710 scope.go:117] "RemoveContainer" containerID="cf79824db6fe9cd2b13ea05cc96c32838c6226bfeb29d1e1acf00214650fe939" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.466034 4710 scope.go:117] "RemoveContainer" containerID="3eae176180b2ecda618a4b9bcd0edcad5d1477673b15adef6751adb2d2d9d8fc" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.526600 4710 scope.go:117] "RemoveContainer" containerID="0add1d91184851a1fef685997472e3d576979615052d37cefd689ec6c2d31e52" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.565152 4710 scope.go:117] "RemoveContainer" containerID="63cd89d4d3711584f92f2146a20d5616ce5357f11db77e97d9dbff50ba00aa3a" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.641168 4710 scope.go:117] "RemoveContainer" containerID="347b532dea3695b2c69836fe68848317a286ff6c35e596e623921f1058c51b2c" Oct 02 13:20:46 crc kubenswrapper[4710]: I1002 13:20:46.677343 4710 scope.go:117] "RemoveContainer" containerID="05b6f86e9266f0e71fffb41fd202dfcc2bc3f3c463e28376d6127873ce718b37" Oct 02 13:20:47 crc kubenswrapper[4710]: I1002 13:20:47.904257 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:20:47 crc kubenswrapper[4710]: E1002 13:20:47.904693 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:20:59 crc kubenswrapper[4710]: I1002 13:20:59.344911 4710 generic.go:334] "Generic (PLEG): container finished" podID="ef9d9d20-85b8-46c0-9696-e05d43c16d39" containerID="7ba02f36a5defcdccabb7fc2f1e5c1ec17857fc787bba554617416cb643ae837" exitCode=0 Oct 02 13:20:59 crc kubenswrapper[4710]: I1002 13:20:59.345002 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" event={"ID":"ef9d9d20-85b8-46c0-9696-e05d43c16d39","Type":"ContainerDied","Data":"7ba02f36a5defcdccabb7fc2f1e5c1ec17857fc787bba554617416cb643ae837"} Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.053809 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fn4fw"] Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.062003 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fn4fw"] Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.783203 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.914716 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32dda3a2-0573-4560-8303-8edc0838c7b9" path="/var/lib/kubelet/pods/32dda3a2-0573-4560-8303-8edc0838c7b9/volumes" Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.956632 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxz8j\" (UniqueName: \"kubernetes.io/projected/ef9d9d20-85b8-46c0-9696-e05d43c16d39-kube-api-access-qxz8j\") pod \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.956771 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-inventory\") pod \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.956836 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-ssh-key\") pod \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\" (UID: \"ef9d9d20-85b8-46c0-9696-e05d43c16d39\") " Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.962966 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9d9d20-85b8-46c0-9696-e05d43c16d39-kube-api-access-qxz8j" (OuterVolumeSpecName: "kube-api-access-qxz8j") pod "ef9d9d20-85b8-46c0-9696-e05d43c16d39" (UID: "ef9d9d20-85b8-46c0-9696-e05d43c16d39"). InnerVolumeSpecName "kube-api-access-qxz8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.986116 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef9d9d20-85b8-46c0-9696-e05d43c16d39" (UID: "ef9d9d20-85b8-46c0-9696-e05d43c16d39"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:21:00 crc kubenswrapper[4710]: I1002 13:21:00.991526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-inventory" (OuterVolumeSpecName: "inventory") pod "ef9d9d20-85b8-46c0-9696-e05d43c16d39" (UID: "ef9d9d20-85b8-46c0-9696-e05d43c16d39"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.059469 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxz8j\" (UniqueName: \"kubernetes.io/projected/ef9d9d20-85b8-46c0-9696-e05d43c16d39-kube-api-access-qxz8j\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.059506 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.059517 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef9d9d20-85b8-46c0-9696-e05d43c16d39-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.365600 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" event={"ID":"ef9d9d20-85b8-46c0-9696-e05d43c16d39","Type":"ContainerDied","Data":"66619cdee2ffc510334cd78b573a03425b5e9bb8346e8ac027d2d39a27caf15e"} Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.365652 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66619cdee2ffc510334cd78b573a03425b5e9bb8346e8ac027d2d39a27caf15e" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.365712 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nf64n" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.450656 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql"] Oct 02 13:21:01 crc kubenswrapper[4710]: E1002 13:21:01.451347 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9d9d20-85b8-46c0-9696-e05d43c16d39" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.451366 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9d9d20-85b8-46c0-9696-e05d43c16d39" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.451712 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9d9d20-85b8-46c0-9696-e05d43c16d39" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.452420 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.464630 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql"] Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.465359 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.465411 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.465697 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.466033 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.575896 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.576444 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.576524 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrvmk\" (UniqueName: \"kubernetes.io/projected/d234b8ec-5b0f-4032-9798-517c5d275c1c-kube-api-access-xrvmk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.679195 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.679273 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrvmk\" (UniqueName: \"kubernetes.io/projected/d234b8ec-5b0f-4032-9798-517c5d275c1c-kube-api-access-xrvmk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.679343 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.685033 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.685071 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.695482 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrvmk\" (UniqueName: \"kubernetes.io/projected/d234b8ec-5b0f-4032-9798-517c5d275c1c-kube-api-access-xrvmk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zrgql\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:01 crc kubenswrapper[4710]: I1002 13:21:01.782403 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:02 crc kubenswrapper[4710]: I1002 13:21:02.280323 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql"] Oct 02 13:21:02 crc kubenswrapper[4710]: I1002 13:21:02.376282 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" event={"ID":"d234b8ec-5b0f-4032-9798-517c5d275c1c","Type":"ContainerStarted","Data":"f4abd316cbe6024045c9b94a6cf979e5de2a4bf4be70bfea7e808b30d44dace4"} Oct 02 13:21:02 crc kubenswrapper[4710]: I1002 13:21:02.905527 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:21:02 crc kubenswrapper[4710]: E1002 13:21:02.906024 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:21:03 crc kubenswrapper[4710]: I1002 13:21:03.385736 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" event={"ID":"d234b8ec-5b0f-4032-9798-517c5d275c1c","Type":"ContainerStarted","Data":"6113c6a25371772ab05c4b550caa5af1b45dae2df4e5ece0d2da9e00c6d56059"} Oct 02 13:21:03 crc kubenswrapper[4710]: I1002 13:21:03.412186 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" podStartSLOduration=1.849447344 podStartE2EDuration="2.412162958s" podCreationTimestamp="2025-10-02 13:21:01 +0000 UTC" firstStartedPulling="2025-10-02 13:21:02.289722784 +0000 UTC m=+1906.396133667" lastFinishedPulling="2025-10-02 13:21:02.852438398 +0000 UTC m=+1906.958849281" observedRunningTime="2025-10-02 13:21:03.400966945 +0000 UTC m=+1907.507377828" watchObservedRunningTime="2025-10-02 13:21:03.412162958 +0000 UTC m=+1907.518573851" Oct 02 13:21:04 crc kubenswrapper[4710]: I1002 13:21:04.032812 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djnzg"] Oct 02 13:21:04 crc kubenswrapper[4710]: I1002 13:21:04.048243 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djnzg"] Oct 02 13:21:04 crc kubenswrapper[4710]: I1002 13:21:04.918933 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d98a36-dd9d-42b5-b26c-a6fa30189c44" path="/var/lib/kubelet/pods/21d98a36-dd9d-42b5-b26c-a6fa30189c44/volumes" Oct 02 13:21:08 crc kubenswrapper[4710]: I1002 13:21:08.429945 4710 generic.go:334] "Generic (PLEG): container finished" podID="d234b8ec-5b0f-4032-9798-517c5d275c1c" containerID="6113c6a25371772ab05c4b550caa5af1b45dae2df4e5ece0d2da9e00c6d56059" exitCode=0 Oct 02 13:21:08 crc kubenswrapper[4710]: I1002 13:21:08.430108 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" event={"ID":"d234b8ec-5b0f-4032-9798-517c5d275c1c","Type":"ContainerDied","Data":"6113c6a25371772ab05c4b550caa5af1b45dae2df4e5ece0d2da9e00c6d56059"} Oct 02 13:21:09 crc kubenswrapper[4710]: I1002 13:21:09.974083 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.171990 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-inventory\") pod \"d234b8ec-5b0f-4032-9798-517c5d275c1c\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.172053 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-ssh-key\") pod \"d234b8ec-5b0f-4032-9798-517c5d275c1c\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.172229 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrvmk\" (UniqueName: \"kubernetes.io/projected/d234b8ec-5b0f-4032-9798-517c5d275c1c-kube-api-access-xrvmk\") pod \"d234b8ec-5b0f-4032-9798-517c5d275c1c\" (UID: \"d234b8ec-5b0f-4032-9798-517c5d275c1c\") " Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.182541 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d234b8ec-5b0f-4032-9798-517c5d275c1c-kube-api-access-xrvmk" (OuterVolumeSpecName: "kube-api-access-xrvmk") pod "d234b8ec-5b0f-4032-9798-517c5d275c1c" (UID: "d234b8ec-5b0f-4032-9798-517c5d275c1c"). InnerVolumeSpecName "kube-api-access-xrvmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.202682 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d234b8ec-5b0f-4032-9798-517c5d275c1c" (UID: "d234b8ec-5b0f-4032-9798-517c5d275c1c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.216529 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-inventory" (OuterVolumeSpecName: "inventory") pod "d234b8ec-5b0f-4032-9798-517c5d275c1c" (UID: "d234b8ec-5b0f-4032-9798-517c5d275c1c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.278222 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.278256 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d234b8ec-5b0f-4032-9798-517c5d275c1c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.278268 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrvmk\" (UniqueName: \"kubernetes.io/projected/d234b8ec-5b0f-4032-9798-517c5d275c1c-kube-api-access-xrvmk\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.449972 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" event={"ID":"d234b8ec-5b0f-4032-9798-517c5d275c1c","Type":"ContainerDied","Data":"f4abd316cbe6024045c9b94a6cf979e5de2a4bf4be70bfea7e808b30d44dace4"} Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.450027 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4abd316cbe6024045c9b94a6cf979e5de2a4bf4be70bfea7e808b30d44dace4" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.450057 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zrgql" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.539130 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g"] Oct 02 13:21:10 crc kubenswrapper[4710]: E1002 13:21:10.539648 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d234b8ec-5b0f-4032-9798-517c5d275c1c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.539676 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d234b8ec-5b0f-4032-9798-517c5d275c1c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.539990 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="d234b8ec-5b0f-4032-9798-517c5d275c1c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.540939 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.546461 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.546923 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.547114 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.547679 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.553770 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g"] Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.586614 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvx2p\" (UniqueName: \"kubernetes.io/projected/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-kube-api-access-tvx2p\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.586673 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.587347 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.689524 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.689698 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvx2p\" (UniqueName: \"kubernetes.io/projected/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-kube-api-access-tvx2p\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.689848 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.694251 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.696392 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.718615 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvx2p\" (UniqueName: \"kubernetes.io/projected/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-kube-api-access-tvx2p\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-t9x9g\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:10 crc kubenswrapper[4710]: I1002 13:21:10.859242 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:11 crc kubenswrapper[4710]: I1002 13:21:11.357275 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g"] Oct 02 13:21:11 crc kubenswrapper[4710]: W1002 13:21:11.361971 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a57743e_4e6c_47f4_8a5f_6a78a4c9b6d9.slice/crio-5d9365479789341629ca914d14e06bd1c94cf5c2fbfdc40642ec4ee3e5e069d0 WatchSource:0}: Error finding container 5d9365479789341629ca914d14e06bd1c94cf5c2fbfdc40642ec4ee3e5e069d0: Status 404 returned error can't find the container with id 5d9365479789341629ca914d14e06bd1c94cf5c2fbfdc40642ec4ee3e5e069d0 Oct 02 13:21:11 crc kubenswrapper[4710]: I1002 13:21:11.469262 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" event={"ID":"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9","Type":"ContainerStarted","Data":"5d9365479789341629ca914d14e06bd1c94cf5c2fbfdc40642ec4ee3e5e069d0"} Oct 02 13:21:12 crc kubenswrapper[4710]: I1002 13:21:12.478919 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" event={"ID":"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9","Type":"ContainerStarted","Data":"ad881c34ac0e6fcb35736df9661a537c67d8b0820dfdaeee235573e158f70e5a"} Oct 02 13:21:12 crc kubenswrapper[4710]: I1002 13:21:12.500862 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" podStartSLOduration=2.075070873 podStartE2EDuration="2.500842236s" podCreationTimestamp="2025-10-02 13:21:10 +0000 UTC" firstStartedPulling="2025-10-02 13:21:11.363887016 +0000 UTC m=+1915.470297899" lastFinishedPulling="2025-10-02 13:21:11.789658379 +0000 UTC m=+1915.896069262" observedRunningTime="2025-10-02 13:21:12.496095167 +0000 UTC m=+1916.602506050" watchObservedRunningTime="2025-10-02 13:21:12.500842236 +0000 UTC m=+1916.607253119" Oct 02 13:21:17 crc kubenswrapper[4710]: I1002 13:21:17.904388 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:21:17 crc kubenswrapper[4710]: E1002 13:21:17.905449 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:21:32 crc kubenswrapper[4710]: I1002 13:21:32.908051 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:21:33 crc kubenswrapper[4710]: I1002 13:21:33.686475 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"e357d2ece9bd04428b3e048314016c37411ee36b3217c0372dc5abe616af8671"} Oct 02 13:21:46 crc kubenswrapper[4710]: I1002 13:21:46.038563 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-grpct"] Oct 02 13:21:46 crc kubenswrapper[4710]: I1002 13:21:46.047681 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-grpct"] Oct 02 13:21:46 crc kubenswrapper[4710]: I1002 13:21:46.855120 4710 scope.go:117] "RemoveContainer" containerID="0160887fdcb5d2cfdf4ca31b2bdcebc3b9ad92ad0f7c4375a93527f91ee161a9" Oct 02 13:21:46 crc kubenswrapper[4710]: I1002 13:21:46.884956 4710 scope.go:117] "RemoveContainer" containerID="c1dc86d76c795049c88be02c4b18e3c9223d8331f1fd017439982d2b75df666b" Oct 02 13:21:46 crc kubenswrapper[4710]: I1002 13:21:46.917114 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cc732de-8466-4872-a4ce-7fc91a64dd8d" path="/var/lib/kubelet/pods/9cc732de-8466-4872-a4ce-7fc91a64dd8d/volumes" Oct 02 13:21:51 crc kubenswrapper[4710]: I1002 13:21:51.849244 4710 generic.go:334] "Generic (PLEG): container finished" podID="3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" containerID="ad881c34ac0e6fcb35736df9661a537c67d8b0820dfdaeee235573e158f70e5a" exitCode=0 Oct 02 13:21:51 crc kubenswrapper[4710]: I1002 13:21:51.849772 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" event={"ID":"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9","Type":"ContainerDied","Data":"ad881c34ac0e6fcb35736df9661a537c67d8b0820dfdaeee235573e158f70e5a"} Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.315795 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.399709 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-inventory\") pod \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.400147 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvx2p\" (UniqueName: \"kubernetes.io/projected/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-kube-api-access-tvx2p\") pod \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.400235 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-ssh-key\") pod \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\" (UID: \"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9\") " Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.405168 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-kube-api-access-tvx2p" (OuterVolumeSpecName: "kube-api-access-tvx2p") pod "3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" (UID: "3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9"). InnerVolumeSpecName "kube-api-access-tvx2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.429899 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" (UID: "3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.433117 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-inventory" (OuterVolumeSpecName: "inventory") pod "3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" (UID: "3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.503364 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.503410 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvx2p\" (UniqueName: \"kubernetes.io/projected/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-kube-api-access-tvx2p\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.503427 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.870866 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" event={"ID":"3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9","Type":"ContainerDied","Data":"5d9365479789341629ca914d14e06bd1c94cf5c2fbfdc40642ec4ee3e5e069d0"} Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.870912 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d9365479789341629ca914d14e06bd1c94cf5c2fbfdc40642ec4ee3e5e069d0" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.870938 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-t9x9g" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.955004 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6"] Oct 02 13:21:53 crc kubenswrapper[4710]: E1002 13:21:53.955473 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.955497 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.955789 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.957517 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.959884 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.960257 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.961420 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.963027 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:21:53 crc kubenswrapper[4710]: I1002 13:21:53.966125 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6"] Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.113978 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.114195 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b2lp\" (UniqueName: \"kubernetes.io/projected/34701c08-ad93-44d4-b39c-1d153bb5f4be-kube-api-access-7b2lp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.114257 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.216419 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.216481 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b2lp\" (UniqueName: \"kubernetes.io/projected/34701c08-ad93-44d4-b39c-1d153bb5f4be-kube-api-access-7b2lp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.216541 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.220062 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.223406 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.235816 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b2lp\" (UniqueName: \"kubernetes.io/projected/34701c08-ad93-44d4-b39c-1d153bb5f4be-kube-api-access-7b2lp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.276472 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.836330 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6"] Oct 02 13:21:54 crc kubenswrapper[4710]: I1002 13:21:54.882318 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" event={"ID":"34701c08-ad93-44d4-b39c-1d153bb5f4be","Type":"ContainerStarted","Data":"e28ddfb894ddd0faf7e34cb38d7293be1fb59fcb767e315ca4520d500c072594"} Oct 02 13:21:55 crc kubenswrapper[4710]: I1002 13:21:55.890702 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" event={"ID":"34701c08-ad93-44d4-b39c-1d153bb5f4be","Type":"ContainerStarted","Data":"430c02c100b0573a2546bd41ff572c81a79cf0b6aa3b85fb8fcfce600e9017c8"} Oct 02 13:21:55 crc kubenswrapper[4710]: I1002 13:21:55.929463 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" podStartSLOduration=2.256791121 podStartE2EDuration="2.929444368s" podCreationTimestamp="2025-10-02 13:21:53 +0000 UTC" firstStartedPulling="2025-10-02 13:21:54.834944427 +0000 UTC m=+1958.941355310" lastFinishedPulling="2025-10-02 13:21:55.507597674 +0000 UTC m=+1959.614008557" observedRunningTime="2025-10-02 13:21:55.91682918 +0000 UTC m=+1960.023240073" watchObservedRunningTime="2025-10-02 13:21:55.929444368 +0000 UTC m=+1960.035855251" Oct 02 13:22:47 crc kubenswrapper[4710]: I1002 13:22:47.006659 4710 scope.go:117] "RemoveContainer" containerID="04a80ba2e91c8740f43c25de43298edef39174bff81dbe4ebde05e6beeb84515" Oct 02 13:22:53 crc kubenswrapper[4710]: I1002 13:22:53.438640 4710 generic.go:334] "Generic (PLEG): container finished" podID="34701c08-ad93-44d4-b39c-1d153bb5f4be" containerID="430c02c100b0573a2546bd41ff572c81a79cf0b6aa3b85fb8fcfce600e9017c8" exitCode=2 Oct 02 13:22:53 crc kubenswrapper[4710]: I1002 13:22:53.438721 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" event={"ID":"34701c08-ad93-44d4-b39c-1d153bb5f4be","Type":"ContainerDied","Data":"430c02c100b0573a2546bd41ff572c81a79cf0b6aa3b85fb8fcfce600e9017c8"} Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.878374 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.937962 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b2lp\" (UniqueName: \"kubernetes.io/projected/34701c08-ad93-44d4-b39c-1d153bb5f4be-kube-api-access-7b2lp\") pod \"34701c08-ad93-44d4-b39c-1d153bb5f4be\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.938068 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-ssh-key\") pod \"34701c08-ad93-44d4-b39c-1d153bb5f4be\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.938246 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-inventory\") pod \"34701c08-ad93-44d4-b39c-1d153bb5f4be\" (UID: \"34701c08-ad93-44d4-b39c-1d153bb5f4be\") " Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.944462 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34701c08-ad93-44d4-b39c-1d153bb5f4be-kube-api-access-7b2lp" (OuterVolumeSpecName: "kube-api-access-7b2lp") pod "34701c08-ad93-44d4-b39c-1d153bb5f4be" (UID: "34701c08-ad93-44d4-b39c-1d153bb5f4be"). InnerVolumeSpecName "kube-api-access-7b2lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.966060 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34701c08-ad93-44d4-b39c-1d153bb5f4be" (UID: "34701c08-ad93-44d4-b39c-1d153bb5f4be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:22:54 crc kubenswrapper[4710]: I1002 13:22:54.971474 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-inventory" (OuterVolumeSpecName: "inventory") pod "34701c08-ad93-44d4-b39c-1d153bb5f4be" (UID: "34701c08-ad93-44d4-b39c-1d153bb5f4be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:22:55 crc kubenswrapper[4710]: I1002 13:22:55.040549 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b2lp\" (UniqueName: \"kubernetes.io/projected/34701c08-ad93-44d4-b39c-1d153bb5f4be-kube-api-access-7b2lp\") on node \"crc\" DevicePath \"\"" Oct 02 13:22:55 crc kubenswrapper[4710]: I1002 13:22:55.040588 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:22:55 crc kubenswrapper[4710]: I1002 13:22:55.040601 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34701c08-ad93-44d4-b39c-1d153bb5f4be-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:22:55 crc kubenswrapper[4710]: I1002 13:22:55.457399 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" event={"ID":"34701c08-ad93-44d4-b39c-1d153bb5f4be","Type":"ContainerDied","Data":"e28ddfb894ddd0faf7e34cb38d7293be1fb59fcb767e315ca4520d500c072594"} Oct 02 13:22:55 crc kubenswrapper[4710]: I1002 13:22:55.457705 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e28ddfb894ddd0faf7e34cb38d7293be1fb59fcb767e315ca4520d500c072594" Oct 02 13:22:55 crc kubenswrapper[4710]: I1002 13:22:55.457456 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.031943 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4"] Oct 02 13:23:03 crc kubenswrapper[4710]: E1002 13:23:03.033107 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34701c08-ad93-44d4-b39c-1d153bb5f4be" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.033137 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="34701c08-ad93-44d4-b39c-1d153bb5f4be" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.033418 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="34701c08-ad93-44d4-b39c-1d153bb5f4be" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.034424 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.035958 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.036229 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.036936 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.038591 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.060952 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4"] Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.100985 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xlrh\" (UniqueName: \"kubernetes.io/projected/b1dad802-c615-4903-89dc-3e19785fa2ed-kube-api-access-9xlrh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.101113 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.101137 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.202781 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.202826 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.202919 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xlrh\" (UniqueName: \"kubernetes.io/projected/b1dad802-c615-4903-89dc-3e19785fa2ed-kube-api-access-9xlrh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.208319 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.216402 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.229501 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xlrh\" (UniqueName: \"kubernetes.io/projected/b1dad802-c615-4903-89dc-3e19785fa2ed-kube-api-access-9xlrh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:03 crc kubenswrapper[4710]: I1002 13:23:03.360159 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:04 crc kubenswrapper[4710]: I1002 13:23:04.011830 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4"] Oct 02 13:23:04 crc kubenswrapper[4710]: I1002 13:23:04.539066 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" event={"ID":"b1dad802-c615-4903-89dc-3e19785fa2ed","Type":"ContainerStarted","Data":"1dee9cc43c9a49745210131534353de8116217e94af3ce91fbc19cece3800cf4"} Oct 02 13:23:05 crc kubenswrapper[4710]: I1002 13:23:05.561328 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" event={"ID":"b1dad802-c615-4903-89dc-3e19785fa2ed","Type":"ContainerStarted","Data":"7ee07e58f8d048a82ed36d12fa2751d3f851c743ba1b7c6f0d24d400ba2557ad"} Oct 02 13:23:05 crc kubenswrapper[4710]: I1002 13:23:05.591095 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" podStartSLOduration=2.16404845 podStartE2EDuration="2.591065364s" podCreationTimestamp="2025-10-02 13:23:03 +0000 UTC" firstStartedPulling="2025-10-02 13:23:04.017148989 +0000 UTC m=+2028.123559872" lastFinishedPulling="2025-10-02 13:23:04.444165903 +0000 UTC m=+2028.550576786" observedRunningTime="2025-10-02 13:23:05.581142234 +0000 UTC m=+2029.687553117" watchObservedRunningTime="2025-10-02 13:23:05.591065364 +0000 UTC m=+2029.697476277" Oct 02 13:23:14 crc kubenswrapper[4710]: I1002 13:23:14.970080 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t764f"] Oct 02 13:23:14 crc kubenswrapper[4710]: I1002 13:23:14.974602 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:14 crc kubenswrapper[4710]: I1002 13:23:14.998823 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t764f"] Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.127475 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-utilities\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.127572 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-catalog-content\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.127605 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jszm\" (UniqueName: \"kubernetes.io/projected/9063d007-c2d0-403a-a418-80a0527bc1b6-kube-api-access-4jszm\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.229252 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-catalog-content\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.229305 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jszm\" (UniqueName: \"kubernetes.io/projected/9063d007-c2d0-403a-a418-80a0527bc1b6-kube-api-access-4jszm\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.229455 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-utilities\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.229862 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-utilities\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.229860 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-catalog-content\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.248540 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jszm\" (UniqueName: \"kubernetes.io/projected/9063d007-c2d0-403a-a418-80a0527bc1b6-kube-api-access-4jszm\") pod \"redhat-marketplace-t764f\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.297961 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:15 crc kubenswrapper[4710]: W1002 13:23:15.757241 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9063d007_c2d0_403a_a418_80a0527bc1b6.slice/crio-62f582f7d75647b416751f081a863da6931a7c9801765902cccfd57b5b1b6bdf WatchSource:0}: Error finding container 62f582f7d75647b416751f081a863da6931a7c9801765902cccfd57b5b1b6bdf: Status 404 returned error can't find the container with id 62f582f7d75647b416751f081a863da6931a7c9801765902cccfd57b5b1b6bdf Oct 02 13:23:15 crc kubenswrapper[4710]: I1002 13:23:15.759040 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t764f"] Oct 02 13:23:16 crc kubenswrapper[4710]: I1002 13:23:16.657560 4710 generic.go:334] "Generic (PLEG): container finished" podID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerID="f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101" exitCode=0 Oct 02 13:23:16 crc kubenswrapper[4710]: I1002 13:23:16.657771 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t764f" event={"ID":"9063d007-c2d0-403a-a418-80a0527bc1b6","Type":"ContainerDied","Data":"f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101"} Oct 02 13:23:16 crc kubenswrapper[4710]: I1002 13:23:16.658086 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t764f" event={"ID":"9063d007-c2d0-403a-a418-80a0527bc1b6","Type":"ContainerStarted","Data":"62f582f7d75647b416751f081a863da6931a7c9801765902cccfd57b5b1b6bdf"} Oct 02 13:23:17 crc kubenswrapper[4710]: I1002 13:23:17.668568 4710 generic.go:334] "Generic (PLEG): container finished" podID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerID="8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36" exitCode=0 Oct 02 13:23:17 crc kubenswrapper[4710]: I1002 13:23:17.668623 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t764f" event={"ID":"9063d007-c2d0-403a-a418-80a0527bc1b6","Type":"ContainerDied","Data":"8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36"} Oct 02 13:23:18 crc kubenswrapper[4710]: I1002 13:23:18.679395 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t764f" event={"ID":"9063d007-c2d0-403a-a418-80a0527bc1b6","Type":"ContainerStarted","Data":"4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7"} Oct 02 13:23:18 crc kubenswrapper[4710]: I1002 13:23:18.700573 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t764f" podStartSLOduration=3.056637908 podStartE2EDuration="4.700550208s" podCreationTimestamp="2025-10-02 13:23:14 +0000 UTC" firstStartedPulling="2025-10-02 13:23:16.659954019 +0000 UTC m=+2040.766364902" lastFinishedPulling="2025-10-02 13:23:18.303866309 +0000 UTC m=+2042.410277202" observedRunningTime="2025-10-02 13:23:18.696904076 +0000 UTC m=+2042.803314969" watchObservedRunningTime="2025-10-02 13:23:18.700550208 +0000 UTC m=+2042.806961091" Oct 02 13:23:25 crc kubenswrapper[4710]: I1002 13:23:25.298400 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:25 crc kubenswrapper[4710]: I1002 13:23:25.298962 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:25 crc kubenswrapper[4710]: I1002 13:23:25.354447 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:25 crc kubenswrapper[4710]: I1002 13:23:25.783261 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:25 crc kubenswrapper[4710]: I1002 13:23:25.830668 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t764f"] Oct 02 13:23:27 crc kubenswrapper[4710]: I1002 13:23:27.760180 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t764f" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="registry-server" containerID="cri-o://4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7" gracePeriod=2 Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.243059 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.308248 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jszm\" (UniqueName: \"kubernetes.io/projected/9063d007-c2d0-403a-a418-80a0527bc1b6-kube-api-access-4jszm\") pod \"9063d007-c2d0-403a-a418-80a0527bc1b6\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.308425 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-utilities\") pod \"9063d007-c2d0-403a-a418-80a0527bc1b6\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.308472 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-catalog-content\") pod \"9063d007-c2d0-403a-a418-80a0527bc1b6\" (UID: \"9063d007-c2d0-403a-a418-80a0527bc1b6\") " Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.314739 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9063d007-c2d0-403a-a418-80a0527bc1b6-kube-api-access-4jszm" (OuterVolumeSpecName: "kube-api-access-4jszm") pod "9063d007-c2d0-403a-a418-80a0527bc1b6" (UID: "9063d007-c2d0-403a-a418-80a0527bc1b6"). InnerVolumeSpecName "kube-api-access-4jszm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.320494 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-utilities" (OuterVolumeSpecName: "utilities") pod "9063d007-c2d0-403a-a418-80a0527bc1b6" (UID: "9063d007-c2d0-403a-a418-80a0527bc1b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.327526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9063d007-c2d0-403a-a418-80a0527bc1b6" (UID: "9063d007-c2d0-403a-a418-80a0527bc1b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.410509 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.410542 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9063d007-c2d0-403a-a418-80a0527bc1b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.410559 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jszm\" (UniqueName: \"kubernetes.io/projected/9063d007-c2d0-403a-a418-80a0527bc1b6-kube-api-access-4jszm\") on node \"crc\" DevicePath \"\"" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.772302 4710 generic.go:334] "Generic (PLEG): container finished" podID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerID="4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7" exitCode=0 Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.772346 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t764f" event={"ID":"9063d007-c2d0-403a-a418-80a0527bc1b6","Type":"ContainerDied","Data":"4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7"} Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.772379 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t764f" event={"ID":"9063d007-c2d0-403a-a418-80a0527bc1b6","Type":"ContainerDied","Data":"62f582f7d75647b416751f081a863da6931a7c9801765902cccfd57b5b1b6bdf"} Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.772384 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t764f" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.772398 4710 scope.go:117] "RemoveContainer" containerID="4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.807140 4710 scope.go:117] "RemoveContainer" containerID="8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.814118 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t764f"] Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.827009 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t764f"] Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.832894 4710 scope.go:117] "RemoveContainer" containerID="f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.900941 4710 scope.go:117] "RemoveContainer" containerID="4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7" Oct 02 13:23:28 crc kubenswrapper[4710]: E1002 13:23:28.901650 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7\": container with ID starting with 4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7 not found: ID does not exist" containerID="4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.901688 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7"} err="failed to get container status \"4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7\": rpc error: code = NotFound desc = could not find container \"4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7\": container with ID starting with 4959d08eba2c79b8797fc1b7b2017f7e7dd63e8b19c0d90357c6ba76d62291d7 not found: ID does not exist" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.901713 4710 scope.go:117] "RemoveContainer" containerID="8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36" Oct 02 13:23:28 crc kubenswrapper[4710]: E1002 13:23:28.901969 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36\": container with ID starting with 8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36 not found: ID does not exist" containerID="8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.901993 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36"} err="failed to get container status \"8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36\": rpc error: code = NotFound desc = could not find container \"8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36\": container with ID starting with 8845d7b3b4c7b2925300ef183ad29eff909c2dfca04957c22eae465faf446a36 not found: ID does not exist" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.902011 4710 scope.go:117] "RemoveContainer" containerID="f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101" Oct 02 13:23:28 crc kubenswrapper[4710]: E1002 13:23:28.902460 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101\": container with ID starting with f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101 not found: ID does not exist" containerID="f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.902478 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101"} err="failed to get container status \"f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101\": rpc error: code = NotFound desc = could not find container \"f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101\": container with ID starting with f9e4bf02245aa7cc30405ab08a5fad09e3993348f0cf1deabec21df9094aa101 not found: ID does not exist" Oct 02 13:23:28 crc kubenswrapper[4710]: I1002 13:23:28.918009 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" path="/var/lib/kubelet/pods/9063d007-c2d0-403a-a418-80a0527bc1b6/volumes" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.779567 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kxkkw"] Oct 02 13:23:49 crc kubenswrapper[4710]: E1002 13:23:49.780631 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="registry-server" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.780647 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="registry-server" Oct 02 13:23:49 crc kubenswrapper[4710]: E1002 13:23:49.780671 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="extract-content" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.780678 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="extract-content" Oct 02 13:23:49 crc kubenswrapper[4710]: E1002 13:23:49.780724 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="extract-utilities" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.780732 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="extract-utilities" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.780991 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="9063d007-c2d0-403a-a418-80a0527bc1b6" containerName="registry-server" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.782793 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.794129 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxkkw"] Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.934583 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416a645-9297-4285-a468-e8741689087b-catalog-content\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.934663 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp9dv\" (UniqueName: \"kubernetes.io/projected/4416a645-9297-4285-a468-e8741689087b-kube-api-access-wp9dv\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:49 crc kubenswrapper[4710]: I1002 13:23:49.934760 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416a645-9297-4285-a468-e8741689087b-utilities\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.036911 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416a645-9297-4285-a468-e8741689087b-catalog-content\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.037250 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp9dv\" (UniqueName: \"kubernetes.io/projected/4416a645-9297-4285-a468-e8741689087b-kube-api-access-wp9dv\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.037420 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416a645-9297-4285-a468-e8741689087b-utilities\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.037608 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416a645-9297-4285-a468-e8741689087b-catalog-content\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.037904 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416a645-9297-4285-a468-e8741689087b-utilities\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.059547 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp9dv\" (UniqueName: \"kubernetes.io/projected/4416a645-9297-4285-a468-e8741689087b-kube-api-access-wp9dv\") pod \"redhat-operators-kxkkw\" (UID: \"4416a645-9297-4285-a468-e8741689087b\") " pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.133892 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.652825 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxkkw"] Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.980131 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxkkw" event={"ID":"4416a645-9297-4285-a468-e8741689087b","Type":"ContainerStarted","Data":"74ab2dbd521b4985200296838e3bc7a281ab9cdf4e2658053c5364371b9f351f"} Oct 02 13:23:50 crc kubenswrapper[4710]: I1002 13:23:50.980499 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxkkw" event={"ID":"4416a645-9297-4285-a468-e8741689087b","Type":"ContainerStarted","Data":"31c7ed2c0801627a76d06b93395d291a3f91a7e4579d279f65dbcea19c2ca8a9"} Oct 02 13:23:51 crc kubenswrapper[4710]: I1002 13:23:51.991531 4710 generic.go:334] "Generic (PLEG): container finished" podID="4416a645-9297-4285-a468-e8741689087b" containerID="74ab2dbd521b4985200296838e3bc7a281ab9cdf4e2658053c5364371b9f351f" exitCode=0 Oct 02 13:23:51 crc kubenswrapper[4710]: I1002 13:23:51.991589 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxkkw" event={"ID":"4416a645-9297-4285-a468-e8741689087b","Type":"ContainerDied","Data":"74ab2dbd521b4985200296838e3bc7a281ab9cdf4e2658053c5364371b9f351f"} Oct 02 13:23:52 crc kubenswrapper[4710]: I1002 13:23:52.530597 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:23:52 crc kubenswrapper[4710]: I1002 13:23:52.530680 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:23:54 crc kubenswrapper[4710]: I1002 13:23:54.013277 4710 generic.go:334] "Generic (PLEG): container finished" podID="b1dad802-c615-4903-89dc-3e19785fa2ed" containerID="7ee07e58f8d048a82ed36d12fa2751d3f851c743ba1b7c6f0d24d400ba2557ad" exitCode=0 Oct 02 13:23:54 crc kubenswrapper[4710]: I1002 13:23:54.013361 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" event={"ID":"b1dad802-c615-4903-89dc-3e19785fa2ed","Type":"ContainerDied","Data":"7ee07e58f8d048a82ed36d12fa2751d3f851c743ba1b7c6f0d24d400ba2557ad"} Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.514928 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.661733 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-inventory\") pod \"b1dad802-c615-4903-89dc-3e19785fa2ed\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.661989 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-ssh-key\") pod \"b1dad802-c615-4903-89dc-3e19785fa2ed\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.662039 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xlrh\" (UniqueName: \"kubernetes.io/projected/b1dad802-c615-4903-89dc-3e19785fa2ed-kube-api-access-9xlrh\") pod \"b1dad802-c615-4903-89dc-3e19785fa2ed\" (UID: \"b1dad802-c615-4903-89dc-3e19785fa2ed\") " Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.669112 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1dad802-c615-4903-89dc-3e19785fa2ed-kube-api-access-9xlrh" (OuterVolumeSpecName: "kube-api-access-9xlrh") pod "b1dad802-c615-4903-89dc-3e19785fa2ed" (UID: "b1dad802-c615-4903-89dc-3e19785fa2ed"). InnerVolumeSpecName "kube-api-access-9xlrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.700582 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-inventory" (OuterVolumeSpecName: "inventory") pod "b1dad802-c615-4903-89dc-3e19785fa2ed" (UID: "b1dad802-c615-4903-89dc-3e19785fa2ed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.709647 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1dad802-c615-4903-89dc-3e19785fa2ed" (UID: "b1dad802-c615-4903-89dc-3e19785fa2ed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.764677 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.764732 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xlrh\" (UniqueName: \"kubernetes.io/projected/b1dad802-c615-4903-89dc-3e19785fa2ed-kube-api-access-9xlrh\") on node \"crc\" DevicePath \"\"" Oct 02 13:23:55 crc kubenswrapper[4710]: I1002 13:23:55.764759 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1dad802-c615-4903-89dc-3e19785fa2ed-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.040554 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" event={"ID":"b1dad802-c615-4903-89dc-3e19785fa2ed","Type":"ContainerDied","Data":"1dee9cc43c9a49745210131534353de8116217e94af3ce91fbc19cece3800cf4"} Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.040623 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dee9cc43c9a49745210131534353de8116217e94af3ce91fbc19cece3800cf4" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.040696 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.124947 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mnx5v"] Oct 02 13:23:56 crc kubenswrapper[4710]: E1002 13:23:56.125641 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1dad802-c615-4903-89dc-3e19785fa2ed" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.125660 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1dad802-c615-4903-89dc-3e19785fa2ed" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.125896 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1dad802-c615-4903-89dc-3e19785fa2ed" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.126635 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.129899 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.130142 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.130306 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.130430 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.160300 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mnx5v"] Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.278331 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.278458 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.278538 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ttwn\" (UniqueName: \"kubernetes.io/projected/40227663-ba0e-4abe-877f-1783781a44a1-kube-api-access-6ttwn\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.379895 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ttwn\" (UniqueName: \"kubernetes.io/projected/40227663-ba0e-4abe-877f-1783781a44a1-kube-api-access-6ttwn\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.380035 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.380118 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.386585 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.387230 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.399734 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ttwn\" (UniqueName: \"kubernetes.io/projected/40227663-ba0e-4abe-877f-1783781a44a1-kube-api-access-6ttwn\") pod \"ssh-known-hosts-edpm-deployment-mnx5v\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:23:56 crc kubenswrapper[4710]: I1002 13:23:56.472259 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:24:03 crc kubenswrapper[4710]: I1002 13:24:03.010816 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mnx5v"] Oct 02 13:24:03 crc kubenswrapper[4710]: I1002 13:24:03.110018 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxkkw" event={"ID":"4416a645-9297-4285-a468-e8741689087b","Type":"ContainerStarted","Data":"b2f91a8fa5f41008b013916f98e792d2e003d026e0fb4b071e7a3796d82fb8d6"} Oct 02 13:24:03 crc kubenswrapper[4710]: I1002 13:24:03.112510 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" event={"ID":"40227663-ba0e-4abe-877f-1783781a44a1","Type":"ContainerStarted","Data":"5a0cd5f86e172507a6843dcf869908a4464e98b5ef7f50c28bb9f69db86e4bcc"} Oct 02 13:24:04 crc kubenswrapper[4710]: I1002 13:24:04.130402 4710 generic.go:334] "Generic (PLEG): container finished" podID="4416a645-9297-4285-a468-e8741689087b" containerID="b2f91a8fa5f41008b013916f98e792d2e003d026e0fb4b071e7a3796d82fb8d6" exitCode=0 Oct 02 13:24:04 crc kubenswrapper[4710]: I1002 13:24:04.130482 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxkkw" event={"ID":"4416a645-9297-4285-a468-e8741689087b","Type":"ContainerDied","Data":"b2f91a8fa5f41008b013916f98e792d2e003d026e0fb4b071e7a3796d82fb8d6"} Oct 02 13:24:04 crc kubenswrapper[4710]: I1002 13:24:04.133446 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" event={"ID":"40227663-ba0e-4abe-877f-1783781a44a1","Type":"ContainerStarted","Data":"1abd4adfb128c3cb853db49fba9d9d07706f0431fbfec5a4e9ffbfc52fb82a4e"} Oct 02 13:24:04 crc kubenswrapper[4710]: I1002 13:24:04.170995 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" podStartSLOduration=7.671985586 podStartE2EDuration="8.17096958s" podCreationTimestamp="2025-10-02 13:23:56 +0000 UTC" firstStartedPulling="2025-10-02 13:24:03.022030307 +0000 UTC m=+2087.128441190" lastFinishedPulling="2025-10-02 13:24:03.521014301 +0000 UTC m=+2087.627425184" observedRunningTime="2025-10-02 13:24:04.162433494 +0000 UTC m=+2088.268844417" watchObservedRunningTime="2025-10-02 13:24:04.17096958 +0000 UTC m=+2088.277380483" Oct 02 13:24:05 crc kubenswrapper[4710]: I1002 13:24:05.158151 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxkkw" event={"ID":"4416a645-9297-4285-a468-e8741689087b","Type":"ContainerStarted","Data":"f869ab16bbb14ae4378b0b98afdefc33d98266676cd871f8a63791fa3d905c37"} Oct 02 13:24:05 crc kubenswrapper[4710]: I1002 13:24:05.196108 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kxkkw" podStartSLOduration=3.560159199 podStartE2EDuration="16.196090834s" podCreationTimestamp="2025-10-02 13:23:49 +0000 UTC" firstStartedPulling="2025-10-02 13:23:51.994233394 +0000 UTC m=+2076.100644277" lastFinishedPulling="2025-10-02 13:24:04.630165029 +0000 UTC m=+2088.736575912" observedRunningTime="2025-10-02 13:24:05.191855617 +0000 UTC m=+2089.298266500" watchObservedRunningTime="2025-10-02 13:24:05.196090834 +0000 UTC m=+2089.302501717" Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.134964 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.135937 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.192493 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.272881 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kxkkw" Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.376241 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxkkw"] Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.445302 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tm6bc"] Oct 02 13:24:10 crc kubenswrapper[4710]: I1002 13:24:10.445606 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tm6bc" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="registry-server" containerID="cri-o://e6690db660b4206d7defb2f2f64f4bf7739773f1549a8b3dcb66cd6070666d20" gracePeriod=2 Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.251260 4710 generic.go:334] "Generic (PLEG): container finished" podID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerID="e6690db660b4206d7defb2f2f64f4bf7739773f1549a8b3dcb66cd6070666d20" exitCode=0 Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.251362 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm6bc" event={"ID":"fccaa30c-236a-4f2a-a2de-644399d3679d","Type":"ContainerDied","Data":"e6690db660b4206d7defb2f2f64f4bf7739773f1549a8b3dcb66cd6070666d20"} Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.425511 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.518592 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87pcs\" (UniqueName: \"kubernetes.io/projected/fccaa30c-236a-4f2a-a2de-644399d3679d-kube-api-access-87pcs\") pod \"fccaa30c-236a-4f2a-a2de-644399d3679d\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.518888 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-utilities\") pod \"fccaa30c-236a-4f2a-a2de-644399d3679d\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.518936 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-catalog-content\") pod \"fccaa30c-236a-4f2a-a2de-644399d3679d\" (UID: \"fccaa30c-236a-4f2a-a2de-644399d3679d\") " Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.529849 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-utilities" (OuterVolumeSpecName: "utilities") pod "fccaa30c-236a-4f2a-a2de-644399d3679d" (UID: "fccaa30c-236a-4f2a-a2de-644399d3679d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.533822 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fccaa30c-236a-4f2a-a2de-644399d3679d-kube-api-access-87pcs" (OuterVolumeSpecName: "kube-api-access-87pcs") pod "fccaa30c-236a-4f2a-a2de-644399d3679d" (UID: "fccaa30c-236a-4f2a-a2de-644399d3679d"). InnerVolumeSpecName "kube-api-access-87pcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.621014 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87pcs\" (UniqueName: \"kubernetes.io/projected/fccaa30c-236a-4f2a-a2de-644399d3679d-kube-api-access-87pcs\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.621057 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.710211 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fccaa30c-236a-4f2a-a2de-644399d3679d" (UID: "fccaa30c-236a-4f2a-a2de-644399d3679d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:24:11 crc kubenswrapper[4710]: I1002 13:24:11.723632 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccaa30c-236a-4f2a-a2de-644399d3679d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.263348 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm6bc" event={"ID":"fccaa30c-236a-4f2a-a2de-644399d3679d","Type":"ContainerDied","Data":"85be7e47c784873282881eca6ea1e898b9d21b5de9928d391582be8c921d7d97"} Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.263357 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm6bc" Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.263717 4710 scope.go:117] "RemoveContainer" containerID="e6690db660b4206d7defb2f2f64f4bf7739773f1549a8b3dcb66cd6070666d20" Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.266480 4710 generic.go:334] "Generic (PLEG): container finished" podID="40227663-ba0e-4abe-877f-1783781a44a1" containerID="1abd4adfb128c3cb853db49fba9d9d07706f0431fbfec5a4e9ffbfc52fb82a4e" exitCode=0 Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.266550 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" event={"ID":"40227663-ba0e-4abe-877f-1783781a44a1","Type":"ContainerDied","Data":"1abd4adfb128c3cb853db49fba9d9d07706f0431fbfec5a4e9ffbfc52fb82a4e"} Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.295972 4710 scope.go:117] "RemoveContainer" containerID="1832a4f418c943fe425198f149841158cc423af7792857c87b80fb112ce945d2" Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.356140 4710 scope.go:117] "RemoveContainer" containerID="d36585f44fd80401717018a5c1fc798721d30594718dd2533a2159c68a343b7a" Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.387870 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tm6bc"] Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.402770 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tm6bc"] Oct 02 13:24:12 crc kubenswrapper[4710]: I1002 13:24:12.915647 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" path="/var/lib/kubelet/pods/fccaa30c-236a-4f2a-a2de-644399d3679d/volumes" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.736947 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.860438 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ttwn\" (UniqueName: \"kubernetes.io/projected/40227663-ba0e-4abe-877f-1783781a44a1-kube-api-access-6ttwn\") pod \"40227663-ba0e-4abe-877f-1783781a44a1\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.860475 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-inventory-0\") pod \"40227663-ba0e-4abe-877f-1783781a44a1\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.860685 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-ssh-key-openstack-edpm-ipam\") pod \"40227663-ba0e-4abe-877f-1783781a44a1\" (UID: \"40227663-ba0e-4abe-877f-1783781a44a1\") " Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.867557 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40227663-ba0e-4abe-877f-1783781a44a1-kube-api-access-6ttwn" (OuterVolumeSpecName: "kube-api-access-6ttwn") pod "40227663-ba0e-4abe-877f-1783781a44a1" (UID: "40227663-ba0e-4abe-877f-1783781a44a1"). InnerVolumeSpecName "kube-api-access-6ttwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.894973 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "40227663-ba0e-4abe-877f-1783781a44a1" (UID: "40227663-ba0e-4abe-877f-1783781a44a1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.912469 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "40227663-ba0e-4abe-877f-1783781a44a1" (UID: "40227663-ba0e-4abe-877f-1783781a44a1"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.963549 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ttwn\" (UniqueName: \"kubernetes.io/projected/40227663-ba0e-4abe-877f-1783781a44a1-kube-api-access-6ttwn\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.963664 4710 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:13 crc kubenswrapper[4710]: I1002 13:24:13.963684 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/40227663-ba0e-4abe-877f-1783781a44a1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.298358 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" event={"ID":"40227663-ba0e-4abe-877f-1783781a44a1","Type":"ContainerDied","Data":"5a0cd5f86e172507a6843dcf869908a4464e98b5ef7f50c28bb9f69db86e4bcc"} Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.298408 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a0cd5f86e172507a6843dcf869908a4464e98b5ef7f50c28bb9f69db86e4bcc" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.298486 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mnx5v" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.383608 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t"] Oct 02 13:24:14 crc kubenswrapper[4710]: E1002 13:24:14.384079 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="extract-utilities" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.384102 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="extract-utilities" Oct 02 13:24:14 crc kubenswrapper[4710]: E1002 13:24:14.384141 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="extract-content" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.384150 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="extract-content" Oct 02 13:24:14 crc kubenswrapper[4710]: E1002 13:24:14.384161 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40227663-ba0e-4abe-877f-1783781a44a1" containerName="ssh-known-hosts-edpm-deployment" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.384168 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="40227663-ba0e-4abe-877f-1783781a44a1" containerName="ssh-known-hosts-edpm-deployment" Oct 02 13:24:14 crc kubenswrapper[4710]: E1002 13:24:14.384183 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="registry-server" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.384192 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="registry-server" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.384390 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="fccaa30c-236a-4f2a-a2de-644399d3679d" containerName="registry-server" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.384414 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="40227663-ba0e-4abe-877f-1783781a44a1" containerName="ssh-known-hosts-edpm-deployment" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.385261 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.388947 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.388959 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.388959 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.390358 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.401955 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t"] Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.473941 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.474034 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.474074 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbm6x\" (UniqueName: \"kubernetes.io/projected/95a95ed4-304b-47b9-9468-819007b0690c-kube-api-access-pbm6x\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.575733 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.575813 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.575848 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbm6x\" (UniqueName: \"kubernetes.io/projected/95a95ed4-304b-47b9-9468-819007b0690c-kube-api-access-pbm6x\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.579719 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.580997 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.598668 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbm6x\" (UniqueName: \"kubernetes.io/projected/95a95ed4-304b-47b9-9468-819007b0690c-kube-api-access-pbm6x\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w925t\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:14 crc kubenswrapper[4710]: I1002 13:24:14.716227 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:15 crc kubenswrapper[4710]: I1002 13:24:15.281124 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t"] Oct 02 13:24:15 crc kubenswrapper[4710]: I1002 13:24:15.315258 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" event={"ID":"95a95ed4-304b-47b9-9468-819007b0690c","Type":"ContainerStarted","Data":"a398613ad9675298f0ef0d6706e55db4201feab08ae3efd534e31d2b2c47999a"} Oct 02 13:24:16 crc kubenswrapper[4710]: I1002 13:24:16.337739 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" event={"ID":"95a95ed4-304b-47b9-9468-819007b0690c","Type":"ContainerStarted","Data":"eb1c234f8e26bab633d0af882069c81239c2935cb131ccd09034d13e7a209b03"} Oct 02 13:24:16 crc kubenswrapper[4710]: I1002 13:24:16.361039 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" podStartSLOduration=1.9026727220000001 podStartE2EDuration="2.36101275s" podCreationTimestamp="2025-10-02 13:24:14 +0000 UTC" firstStartedPulling="2025-10-02 13:24:15.291965106 +0000 UTC m=+2099.398375979" lastFinishedPulling="2025-10-02 13:24:15.750305124 +0000 UTC m=+2099.856716007" observedRunningTime="2025-10-02 13:24:16.358540247 +0000 UTC m=+2100.464951140" watchObservedRunningTime="2025-10-02 13:24:16.36101275 +0000 UTC m=+2100.467423633" Oct 02 13:24:22 crc kubenswrapper[4710]: I1002 13:24:22.530827 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:24:22 crc kubenswrapper[4710]: I1002 13:24:22.531380 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:24:25 crc kubenswrapper[4710]: I1002 13:24:25.412084 4710 generic.go:334] "Generic (PLEG): container finished" podID="95a95ed4-304b-47b9-9468-819007b0690c" containerID="eb1c234f8e26bab633d0af882069c81239c2935cb131ccd09034d13e7a209b03" exitCode=0 Oct 02 13:24:25 crc kubenswrapper[4710]: I1002 13:24:25.412171 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" event={"ID":"95a95ed4-304b-47b9-9468-819007b0690c","Type":"ContainerDied","Data":"eb1c234f8e26bab633d0af882069c81239c2935cb131ccd09034d13e7a209b03"} Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.898235 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.919791 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbm6x\" (UniqueName: \"kubernetes.io/projected/95a95ed4-304b-47b9-9468-819007b0690c-kube-api-access-pbm6x\") pod \"95a95ed4-304b-47b9-9468-819007b0690c\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.919881 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-inventory\") pod \"95a95ed4-304b-47b9-9468-819007b0690c\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.920119 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-ssh-key\") pod \"95a95ed4-304b-47b9-9468-819007b0690c\" (UID: \"95a95ed4-304b-47b9-9468-819007b0690c\") " Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.933354 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a95ed4-304b-47b9-9468-819007b0690c-kube-api-access-pbm6x" (OuterVolumeSpecName: "kube-api-access-pbm6x") pod "95a95ed4-304b-47b9-9468-819007b0690c" (UID: "95a95ed4-304b-47b9-9468-819007b0690c"). InnerVolumeSpecName "kube-api-access-pbm6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.955954 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-inventory" (OuterVolumeSpecName: "inventory") pod "95a95ed4-304b-47b9-9468-819007b0690c" (UID: "95a95ed4-304b-47b9-9468-819007b0690c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:24:26 crc kubenswrapper[4710]: I1002 13:24:26.958285 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "95a95ed4-304b-47b9-9468-819007b0690c" (UID: "95a95ed4-304b-47b9-9468-819007b0690c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.022194 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.022231 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbm6x\" (UniqueName: \"kubernetes.io/projected/95a95ed4-304b-47b9-9468-819007b0690c-kube-api-access-pbm6x\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.022245 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a95ed4-304b-47b9-9468-819007b0690c-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.443781 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" event={"ID":"95a95ed4-304b-47b9-9468-819007b0690c","Type":"ContainerDied","Data":"a398613ad9675298f0ef0d6706e55db4201feab08ae3efd534e31d2b2c47999a"} Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.443826 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a398613ad9675298f0ef0d6706e55db4201feab08ae3efd534e31d2b2c47999a" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.443827 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w925t" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.511669 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl"] Oct 02 13:24:27 crc kubenswrapper[4710]: E1002 13:24:27.512267 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a95ed4-304b-47b9-9468-819007b0690c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.512285 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a95ed4-304b-47b9-9468-819007b0690c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.512511 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a95ed4-304b-47b9-9468-819007b0690c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.513259 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.515529 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.515797 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.516196 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.516705 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.520874 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl"] Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.690062 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.690132 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.690220 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skj65\" (UniqueName: \"kubernetes.io/projected/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-kube-api-access-skj65\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.794181 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.794300 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.794500 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skj65\" (UniqueName: \"kubernetes.io/projected/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-kube-api-access-skj65\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.799421 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.801166 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.822604 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skj65\" (UniqueName: \"kubernetes.io/projected/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-kube-api-access-skj65\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:27 crc kubenswrapper[4710]: I1002 13:24:27.898714 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:28 crc kubenswrapper[4710]: I1002 13:24:28.423152 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl"] Oct 02 13:24:28 crc kubenswrapper[4710]: I1002 13:24:28.456790 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" event={"ID":"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5","Type":"ContainerStarted","Data":"3191b7293689a824c6c4f4f36824fafeee119eabb0eb42c4e6abd55252213842"} Oct 02 13:24:29 crc kubenswrapper[4710]: I1002 13:24:29.468223 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" event={"ID":"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5","Type":"ContainerStarted","Data":"7be65af91faa7241f5a60fccee20d3985fb57a722c9cf0d7e82cddcce184a727"} Oct 02 13:24:29 crc kubenswrapper[4710]: I1002 13:24:29.500344 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" podStartSLOduration=2.060199802 podStartE2EDuration="2.50031682s" podCreationTimestamp="2025-10-02 13:24:27 +0000 UTC" firstStartedPulling="2025-10-02 13:24:28.436152149 +0000 UTC m=+2112.542563032" lastFinishedPulling="2025-10-02 13:24:28.876269157 +0000 UTC m=+2112.982680050" observedRunningTime="2025-10-02 13:24:29.484736077 +0000 UTC m=+2113.591146960" watchObservedRunningTime="2025-10-02 13:24:29.50031682 +0000 UTC m=+2113.606727723" Oct 02 13:24:39 crc kubenswrapper[4710]: I1002 13:24:39.565124 4710 generic.go:334] "Generic (PLEG): container finished" podID="7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" containerID="7be65af91faa7241f5a60fccee20d3985fb57a722c9cf0d7e82cddcce184a727" exitCode=0 Oct 02 13:24:39 crc kubenswrapper[4710]: I1002 13:24:39.565259 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" event={"ID":"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5","Type":"ContainerDied","Data":"7be65af91faa7241f5a60fccee20d3985fb57a722c9cf0d7e82cddcce184a727"} Oct 02 13:24:40 crc kubenswrapper[4710]: I1002 13:24:40.930140 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.073487 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-ssh-key\") pod \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.073565 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skj65\" (UniqueName: \"kubernetes.io/projected/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-kube-api-access-skj65\") pod \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.073643 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-inventory\") pod \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\" (UID: \"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5\") " Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.105191 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-kube-api-access-skj65" (OuterVolumeSpecName: "kube-api-access-skj65") pod "7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" (UID: "7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5"). InnerVolumeSpecName "kube-api-access-skj65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.110341 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-inventory" (OuterVolumeSpecName: "inventory") pod "7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" (UID: "7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.115742 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" (UID: "7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.178183 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.178247 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skj65\" (UniqueName: \"kubernetes.io/projected/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-kube-api-access-skj65\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.178279 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.585668 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" event={"ID":"7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5","Type":"ContainerDied","Data":"3191b7293689a824c6c4f4f36824fafeee119eabb0eb42c4e6abd55252213842"} Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.586012 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3191b7293689a824c6c4f4f36824fafeee119eabb0eb42c4e6abd55252213842" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.585728 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.684635 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz"] Oct 02 13:24:41 crc kubenswrapper[4710]: E1002 13:24:41.685189 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.685211 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.685465 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.687110 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.689574 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.689899 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.690099 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.690233 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.690331 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.690373 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.690557 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.692412 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.695621 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz"] Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.791455 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.791537 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.791696 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwcdb\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-kube-api-access-nwcdb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792092 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792323 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792454 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792544 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792629 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792688 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792775 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792813 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.792978 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.793043 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.793160 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.894658 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.895635 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.895729 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.895815 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896180 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896226 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896294 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896376 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896401 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896438 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896482 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896509 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896554 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwcdb\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-kube-api-access-nwcdb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.896580 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.900240 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.901046 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.901691 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.901878 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.902408 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.902708 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.902823 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.902831 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.904324 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.906212 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.906222 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.906672 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.911129 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:41 crc kubenswrapper[4710]: I1002 13:24:41.913949 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwcdb\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-kube-api-access-nwcdb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:42 crc kubenswrapper[4710]: I1002 13:24:42.013439 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:24:42 crc kubenswrapper[4710]: I1002 13:24:42.587994 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz"] Oct 02 13:24:42 crc kubenswrapper[4710]: I1002 13:24:42.590387 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:24:43 crc kubenswrapper[4710]: I1002 13:24:43.603509 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" event={"ID":"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed","Type":"ContainerStarted","Data":"2cb18bae6c93b7a3c53b13ce4fc25ba9bdfd91acc4d27b019a65a234aa6a852d"} Oct 02 13:24:44 crc kubenswrapper[4710]: I1002 13:24:44.613302 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" event={"ID":"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed","Type":"ContainerStarted","Data":"a6672b6d541e0ea5fee6f613f08b2ba0b81ca96b91d633a90bf6eaf739403af6"} Oct 02 13:24:44 crc kubenswrapper[4710]: I1002 13:24:44.647660 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" podStartSLOduration=2.883377457 podStartE2EDuration="3.647640982s" podCreationTimestamp="2025-10-02 13:24:41 +0000 UTC" firstStartedPulling="2025-10-02 13:24:42.590206952 +0000 UTC m=+2126.696617835" lastFinishedPulling="2025-10-02 13:24:43.354470477 +0000 UTC m=+2127.460881360" observedRunningTime="2025-10-02 13:24:44.643395305 +0000 UTC m=+2128.749806208" watchObservedRunningTime="2025-10-02 13:24:44.647640982 +0000 UTC m=+2128.754051855" Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.530546 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.531179 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.531236 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.532226 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e357d2ece9bd04428b3e048314016c37411ee36b3217c0372dc5abe616af8671"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.532319 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://e357d2ece9bd04428b3e048314016c37411ee36b3217c0372dc5abe616af8671" gracePeriod=600 Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.731131 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="e357d2ece9bd04428b3e048314016c37411ee36b3217c0372dc5abe616af8671" exitCode=0 Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.731217 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"e357d2ece9bd04428b3e048314016c37411ee36b3217c0372dc5abe616af8671"} Oct 02 13:24:52 crc kubenswrapper[4710]: I1002 13:24:52.731451 4710 scope.go:117] "RemoveContainer" containerID="24a510490f970e7858c2080d73a089119706199d8d049b5c581c6bb52f49052e" Oct 02 13:24:53 crc kubenswrapper[4710]: I1002 13:24:53.742924 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801"} Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.070687 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hrhmw"] Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.073648 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.097203 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrhmw"] Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.259536 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-utilities\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.259603 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttvmj\" (UniqueName: \"kubernetes.io/projected/b9567df0-094c-45ff-9c19-c1d34b57b5ff-kube-api-access-ttvmj\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.259668 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-catalog-content\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.362054 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-utilities\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.362135 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttvmj\" (UniqueName: \"kubernetes.io/projected/b9567df0-094c-45ff-9c19-c1d34b57b5ff-kube-api-access-ttvmj\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.362202 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-catalog-content\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.362594 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-utilities\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.362645 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-catalog-content\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.387947 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttvmj\" (UniqueName: \"kubernetes.io/projected/b9567df0-094c-45ff-9c19-c1d34b57b5ff-kube-api-access-ttvmj\") pod \"certified-operators-hrhmw\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.402947 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:24:54 crc kubenswrapper[4710]: I1002 13:24:54.960378 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrhmw"] Oct 02 13:24:55 crc kubenswrapper[4710]: I1002 13:24:55.771684 4710 generic.go:334] "Generic (PLEG): container finished" podID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerID="8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b" exitCode=0 Oct 02 13:24:55 crc kubenswrapper[4710]: I1002 13:24:55.772396 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerDied","Data":"8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b"} Oct 02 13:24:55 crc kubenswrapper[4710]: I1002 13:24:55.772472 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerStarted","Data":"2bc54fab5d570dc7d5384e2dafb189afef35da461738dbfc2904febf9b3f4103"} Oct 02 13:24:57 crc kubenswrapper[4710]: I1002 13:24:57.794149 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerStarted","Data":"f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71"} Oct 02 13:24:58 crc kubenswrapper[4710]: I1002 13:24:58.806589 4710 generic.go:334] "Generic (PLEG): container finished" podID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerID="f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71" exitCode=0 Oct 02 13:24:58 crc kubenswrapper[4710]: I1002 13:24:58.806653 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerDied","Data":"f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71"} Oct 02 13:24:59 crc kubenswrapper[4710]: I1002 13:24:59.819117 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerStarted","Data":"76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26"} Oct 02 13:24:59 crc kubenswrapper[4710]: I1002 13:24:59.849278 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hrhmw" podStartSLOduration=2.179314683 podStartE2EDuration="5.849259566s" podCreationTimestamp="2025-10-02 13:24:54 +0000 UTC" firstStartedPulling="2025-10-02 13:24:55.777315658 +0000 UTC m=+2139.883726561" lastFinishedPulling="2025-10-02 13:24:59.447260561 +0000 UTC m=+2143.553671444" observedRunningTime="2025-10-02 13:24:59.8427078 +0000 UTC m=+2143.949118703" watchObservedRunningTime="2025-10-02 13:24:59.849259566 +0000 UTC m=+2143.955670449" Oct 02 13:25:04 crc kubenswrapper[4710]: I1002 13:25:04.404593 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:25:04 crc kubenswrapper[4710]: I1002 13:25:04.405176 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:25:04 crc kubenswrapper[4710]: I1002 13:25:04.457648 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:25:04 crc kubenswrapper[4710]: I1002 13:25:04.923497 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:25:04 crc kubenswrapper[4710]: I1002 13:25:04.972941 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrhmw"] Oct 02 13:25:06 crc kubenswrapper[4710]: I1002 13:25:06.885364 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hrhmw" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="registry-server" containerID="cri-o://76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26" gracePeriod=2 Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.358199 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.554773 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-utilities\") pod \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.554961 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttvmj\" (UniqueName: \"kubernetes.io/projected/b9567df0-094c-45ff-9c19-c1d34b57b5ff-kube-api-access-ttvmj\") pod \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.555040 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-catalog-content\") pod \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\" (UID: \"b9567df0-094c-45ff-9c19-c1d34b57b5ff\") " Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.555984 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-utilities" (OuterVolumeSpecName: "utilities") pod "b9567df0-094c-45ff-9c19-c1d34b57b5ff" (UID: "b9567df0-094c-45ff-9c19-c1d34b57b5ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.560270 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9567df0-094c-45ff-9c19-c1d34b57b5ff-kube-api-access-ttvmj" (OuterVolumeSpecName: "kube-api-access-ttvmj") pod "b9567df0-094c-45ff-9c19-c1d34b57b5ff" (UID: "b9567df0-094c-45ff-9c19-c1d34b57b5ff"). InnerVolumeSpecName "kube-api-access-ttvmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.608045 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9567df0-094c-45ff-9c19-c1d34b57b5ff" (UID: "b9567df0-094c-45ff-9c19-c1d34b57b5ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.659052 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.659098 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttvmj\" (UniqueName: \"kubernetes.io/projected/b9567df0-094c-45ff-9c19-c1d34b57b5ff-kube-api-access-ttvmj\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.659117 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9567df0-094c-45ff-9c19-c1d34b57b5ff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.898175 4710 generic.go:334] "Generic (PLEG): container finished" podID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerID="76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26" exitCode=0 Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.898244 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerDied","Data":"76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26"} Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.898286 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhmw" event={"ID":"b9567df0-094c-45ff-9c19-c1d34b57b5ff","Type":"ContainerDied","Data":"2bc54fab5d570dc7d5384e2dafb189afef35da461738dbfc2904febf9b3f4103"} Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.898308 4710 scope.go:117] "RemoveContainer" containerID="76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.898392 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhmw" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.931101 4710 scope.go:117] "RemoveContainer" containerID="f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71" Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.943773 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrhmw"] Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.953838 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hrhmw"] Oct 02 13:25:07 crc kubenswrapper[4710]: I1002 13:25:07.975295 4710 scope.go:117] "RemoveContainer" containerID="8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.012988 4710 scope.go:117] "RemoveContainer" containerID="76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26" Oct 02 13:25:08 crc kubenswrapper[4710]: E1002 13:25:08.013513 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26\": container with ID starting with 76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26 not found: ID does not exist" containerID="76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.013582 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26"} err="failed to get container status \"76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26\": rpc error: code = NotFound desc = could not find container \"76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26\": container with ID starting with 76d60288c3df6eb9c99f8a9ba914d4b3b067bda7fce4dfd244f4af56c6648b26 not found: ID does not exist" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.013615 4710 scope.go:117] "RemoveContainer" containerID="f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71" Oct 02 13:25:08 crc kubenswrapper[4710]: E1002 13:25:08.014096 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71\": container with ID starting with f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71 not found: ID does not exist" containerID="f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.014137 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71"} err="failed to get container status \"f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71\": rpc error: code = NotFound desc = could not find container \"f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71\": container with ID starting with f1021fc183a74bce472298aada3535a2fe46ec1ce5dc9ff2ac1f209852370c71 not found: ID does not exist" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.014170 4710 scope.go:117] "RemoveContainer" containerID="8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b" Oct 02 13:25:08 crc kubenswrapper[4710]: E1002 13:25:08.014450 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b\": container with ID starting with 8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b not found: ID does not exist" containerID="8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.014490 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b"} err="failed to get container status \"8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b\": rpc error: code = NotFound desc = could not find container \"8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b\": container with ID starting with 8dc1153e5acefbea9e510c04d05af8d6597bc04c9a0fa77ec18ce1a82fad495b not found: ID does not exist" Oct 02 13:25:08 crc kubenswrapper[4710]: I1002 13:25:08.920123 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" path="/var/lib/kubelet/pods/b9567df0-094c-45ff-9c19-c1d34b57b5ff/volumes" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.442306 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b9dq4"] Oct 02 13:25:17 crc kubenswrapper[4710]: E1002 13:25:17.443777 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="extract-content" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.443805 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="extract-content" Oct 02 13:25:17 crc kubenswrapper[4710]: E1002 13:25:17.443850 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="extract-utilities" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.443864 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="extract-utilities" Oct 02 13:25:17 crc kubenswrapper[4710]: E1002 13:25:17.443889 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="registry-server" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.443901 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="registry-server" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.444239 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9567df0-094c-45ff-9c19-c1d34b57b5ff" containerName="registry-server" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.446799 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.453314 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b9dq4"] Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.560463 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmhv8\" (UniqueName: \"kubernetes.io/projected/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-kube-api-access-bmhv8\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.560583 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-utilities\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.560712 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-catalog-content\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.662512 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-utilities\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.662687 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-catalog-content\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.662811 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmhv8\" (UniqueName: \"kubernetes.io/projected/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-kube-api-access-bmhv8\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.663065 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-utilities\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.663121 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-catalog-content\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.690356 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmhv8\" (UniqueName: \"kubernetes.io/projected/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-kube-api-access-bmhv8\") pod \"community-operators-b9dq4\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:17 crc kubenswrapper[4710]: I1002 13:25:17.765618 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:18 crc kubenswrapper[4710]: I1002 13:25:18.103657 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b9dq4"] Oct 02 13:25:19 crc kubenswrapper[4710]: I1002 13:25:19.024038 4710 generic.go:334] "Generic (PLEG): container finished" podID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerID="8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b" exitCode=0 Oct 02 13:25:19 crc kubenswrapper[4710]: I1002 13:25:19.024355 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerDied","Data":"8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b"} Oct 02 13:25:19 crc kubenswrapper[4710]: I1002 13:25:19.024384 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerStarted","Data":"f7bb0c8f439170789527f828c9a8bdab0d069b56a472b5c972cd976293f67a99"} Oct 02 13:25:20 crc kubenswrapper[4710]: I1002 13:25:20.035483 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerStarted","Data":"30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457"} Oct 02 13:25:24 crc kubenswrapper[4710]: I1002 13:25:24.078318 4710 generic.go:334] "Generic (PLEG): container finished" podID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerID="30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457" exitCode=0 Oct 02 13:25:24 crc kubenswrapper[4710]: I1002 13:25:24.078441 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerDied","Data":"30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457"} Oct 02 13:25:25 crc kubenswrapper[4710]: I1002 13:25:25.090619 4710 generic.go:334] "Generic (PLEG): container finished" podID="c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" containerID="a6672b6d541e0ea5fee6f613f08b2ba0b81ca96b91d633a90bf6eaf739403af6" exitCode=0 Oct 02 13:25:25 crc kubenswrapper[4710]: I1002 13:25:25.090715 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" event={"ID":"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed","Type":"ContainerDied","Data":"a6672b6d541e0ea5fee6f613f08b2ba0b81ca96b91d633a90bf6eaf739403af6"} Oct 02 13:25:25 crc kubenswrapper[4710]: I1002 13:25:25.095808 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerStarted","Data":"1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef"} Oct 02 13:25:25 crc kubenswrapper[4710]: I1002 13:25:25.146384 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b9dq4" podStartSLOduration=2.5928286099999998 podStartE2EDuration="8.146363513s" podCreationTimestamp="2025-10-02 13:25:17 +0000 UTC" firstStartedPulling="2025-10-02 13:25:19.026081984 +0000 UTC m=+2163.132492877" lastFinishedPulling="2025-10-02 13:25:24.579616897 +0000 UTC m=+2168.686027780" observedRunningTime="2025-10-02 13:25:25.138638428 +0000 UTC m=+2169.245049321" watchObservedRunningTime="2025-10-02 13:25:25.146363513 +0000 UTC m=+2169.252774406" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.547810 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.650330 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.650386 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-bootstrap-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.650419 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-neutron-metadata-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651304 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-nova-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651339 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-telemetry-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651372 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-libvirt-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651393 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-inventory\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651457 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ovn-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651476 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651495 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651515 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651541 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-repo-setup-combined-ca-bundle\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651564 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwcdb\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-kube-api-access-nwcdb\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.651605 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ssh-key\") pod \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\" (UID: \"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed\") " Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.658294 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.658461 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.658480 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.659636 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.659673 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.661213 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.661633 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.664160 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-kube-api-access-nwcdb" (OuterVolumeSpecName: "kube-api-access-nwcdb") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "kube-api-access-nwcdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.664403 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.669667 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.670577 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.680314 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.698163 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-inventory" (OuterVolumeSpecName: "inventory") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.714152 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" (UID: "c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.752794 4710 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.753667 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.753799 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.753903 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.753987 4710 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754066 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwcdb\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-kube-api-access-nwcdb\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754138 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754217 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754483 4710 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754589 4710 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754690 4710 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754807 4710 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754884 4710 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:26 crc kubenswrapper[4710]: I1002 13:25:26.754953 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.118164 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" event={"ID":"c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed","Type":"ContainerDied","Data":"2cb18bae6c93b7a3c53b13ce4fc25ba9bdfd91acc4d27b019a65a234aa6a852d"} Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.118463 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cb18bae6c93b7a3c53b13ce4fc25ba9bdfd91acc4d27b019a65a234aa6a852d" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.118240 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.223926 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m"] Oct 02 13:25:27 crc kubenswrapper[4710]: E1002 13:25:27.224417 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.224434 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.224639 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.225438 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.228877 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.228914 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.228957 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.228996 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.229033 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.231825 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m"] Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.363737 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.363816 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.363898 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.364907 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jnn8\" (UniqueName: \"kubernetes.io/projected/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-kube-api-access-4jnn8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.365113 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.466684 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jnn8\" (UniqueName: \"kubernetes.io/projected/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-kube-api-access-4jnn8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.466785 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.466855 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.466903 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.467107 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.467840 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.471721 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.471806 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.478474 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.488850 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jnn8\" (UniqueName: \"kubernetes.io/projected/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-kube-api-access-4jnn8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hcv8m\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.542108 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.766078 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.766374 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:27 crc kubenswrapper[4710]: I1002 13:25:27.819769 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:28 crc kubenswrapper[4710]: I1002 13:25:28.083729 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m"] Oct 02 13:25:28 crc kubenswrapper[4710]: I1002 13:25:28.126378 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" event={"ID":"6c66d4e3-dc39-4b2d-9adf-c3c69b149982","Type":"ContainerStarted","Data":"ea5249ea1fd0aef7d9866a1f5dc1925a56ae07acee7d3b8930d6c12a1b621cee"} Oct 02 13:25:29 crc kubenswrapper[4710]: I1002 13:25:29.136621 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" event={"ID":"6c66d4e3-dc39-4b2d-9adf-c3c69b149982","Type":"ContainerStarted","Data":"4b4101c804fff3ba03814d0c549c10dd62ea9b42fb576e30532ab4aff5585f13"} Oct 02 13:25:29 crc kubenswrapper[4710]: I1002 13:25:29.165638 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" podStartSLOduration=1.7709281300000002 podStartE2EDuration="2.165611469s" podCreationTimestamp="2025-10-02 13:25:27 +0000 UTC" firstStartedPulling="2025-10-02 13:25:28.089715982 +0000 UTC m=+2172.196126855" lastFinishedPulling="2025-10-02 13:25:28.484399301 +0000 UTC m=+2172.590810194" observedRunningTime="2025-10-02 13:25:29.157509774 +0000 UTC m=+2173.263920697" watchObservedRunningTime="2025-10-02 13:25:29.165611469 +0000 UTC m=+2173.272022392" Oct 02 13:25:37 crc kubenswrapper[4710]: I1002 13:25:37.826139 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:37 crc kubenswrapper[4710]: I1002 13:25:37.875122 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b9dq4"] Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.222160 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b9dq4" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="registry-server" containerID="cri-o://1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef" gracePeriod=2 Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.724407 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.892089 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-catalog-content\") pod \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.892195 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-utilities\") pod \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.892245 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmhv8\" (UniqueName: \"kubernetes.io/projected/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-kube-api-access-bmhv8\") pod \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\" (UID: \"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c\") " Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.892856 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-utilities" (OuterVolumeSpecName: "utilities") pod "d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" (UID: "d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.906980 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-kube-api-access-bmhv8" (OuterVolumeSpecName: "kube-api-access-bmhv8") pod "d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" (UID: "d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c"). InnerVolumeSpecName "kube-api-access-bmhv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.942927 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" (UID: "d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.994528 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.994575 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmhv8\" (UniqueName: \"kubernetes.io/projected/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-kube-api-access-bmhv8\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:38 crc kubenswrapper[4710]: I1002 13:25:38.994589 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.232890 4710 generic.go:334] "Generic (PLEG): container finished" podID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerID="1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef" exitCode=0 Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.232931 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9dq4" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.232937 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerDied","Data":"1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef"} Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.232975 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9dq4" event={"ID":"d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c","Type":"ContainerDied","Data":"f7bb0c8f439170789527f828c9a8bdab0d069b56a472b5c972cd976293f67a99"} Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.233003 4710 scope.go:117] "RemoveContainer" containerID="1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.255588 4710 scope.go:117] "RemoveContainer" containerID="30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.276547 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b9dq4"] Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.283799 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b9dq4"] Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.291577 4710 scope.go:117] "RemoveContainer" containerID="8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.328062 4710 scope.go:117] "RemoveContainer" containerID="1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef" Oct 02 13:25:39 crc kubenswrapper[4710]: E1002 13:25:39.328405 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef\": container with ID starting with 1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef not found: ID does not exist" containerID="1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.328438 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef"} err="failed to get container status \"1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef\": rpc error: code = NotFound desc = could not find container \"1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef\": container with ID starting with 1dd4a3fc50e54e022ae327184cfe888c686c6006731faadeded22168fc0cb5ef not found: ID does not exist" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.328456 4710 scope.go:117] "RemoveContainer" containerID="30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457" Oct 02 13:25:39 crc kubenswrapper[4710]: E1002 13:25:39.328645 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457\": container with ID starting with 30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457 not found: ID does not exist" containerID="30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.328665 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457"} err="failed to get container status \"30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457\": rpc error: code = NotFound desc = could not find container \"30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457\": container with ID starting with 30efeb6dc08de5a636801b681e484e428b74cb9df1d22829b37c96d7c2511457 not found: ID does not exist" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.328678 4710 scope.go:117] "RemoveContainer" containerID="8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b" Oct 02 13:25:39 crc kubenswrapper[4710]: E1002 13:25:39.328843 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b\": container with ID starting with 8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b not found: ID does not exist" containerID="8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b" Oct 02 13:25:39 crc kubenswrapper[4710]: I1002 13:25:39.328859 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b"} err="failed to get container status \"8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b\": rpc error: code = NotFound desc = could not find container \"8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b\": container with ID starting with 8f607b0fc72f75168e5eceb60080a5eb3138ae30c27cc9c81d746e253880741b not found: ID does not exist" Oct 02 13:25:40 crc kubenswrapper[4710]: I1002 13:25:40.916366 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" path="/var/lib/kubelet/pods/d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c/volumes" Oct 02 13:26:36 crc kubenswrapper[4710]: I1002 13:26:36.781762 4710 generic.go:334] "Generic (PLEG): container finished" podID="6c66d4e3-dc39-4b2d-9adf-c3c69b149982" containerID="4b4101c804fff3ba03814d0c549c10dd62ea9b42fb576e30532ab4aff5585f13" exitCode=0 Oct 02 13:26:36 crc kubenswrapper[4710]: I1002 13:26:36.781879 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" event={"ID":"6c66d4e3-dc39-4b2d-9adf-c3c69b149982","Type":"ContainerDied","Data":"4b4101c804fff3ba03814d0c549c10dd62ea9b42fb576e30532ab4aff5585f13"} Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.257058 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.407287 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ssh-key\") pod \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.407529 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovncontroller-config-0\") pod \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.407703 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovn-combined-ca-bundle\") pod \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.407766 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-inventory\") pod \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.407912 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jnn8\" (UniqueName: \"kubernetes.io/projected/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-kube-api-access-4jnn8\") pod \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\" (UID: \"6c66d4e3-dc39-4b2d-9adf-c3c69b149982\") " Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.417196 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6c66d4e3-dc39-4b2d-9adf-c3c69b149982" (UID: "6c66d4e3-dc39-4b2d-9adf-c3c69b149982"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.417218 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-kube-api-access-4jnn8" (OuterVolumeSpecName: "kube-api-access-4jnn8") pod "6c66d4e3-dc39-4b2d-9adf-c3c69b149982" (UID: "6c66d4e3-dc39-4b2d-9adf-c3c69b149982"). InnerVolumeSpecName "kube-api-access-4jnn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.442579 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6c66d4e3-dc39-4b2d-9adf-c3c69b149982" (UID: "6c66d4e3-dc39-4b2d-9adf-c3c69b149982"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.442616 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-inventory" (OuterVolumeSpecName: "inventory") pod "6c66d4e3-dc39-4b2d-9adf-c3c69b149982" (UID: "6c66d4e3-dc39-4b2d-9adf-c3c69b149982"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.448583 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6c66d4e3-dc39-4b2d-9adf-c3c69b149982" (UID: "6c66d4e3-dc39-4b2d-9adf-c3c69b149982"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.511866 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.511922 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jnn8\" (UniqueName: \"kubernetes.io/projected/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-kube-api-access-4jnn8\") on node \"crc\" DevicePath \"\"" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.511945 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.511963 4710 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.511981 4710 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c66d4e3-dc39-4b2d-9adf-c3c69b149982-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.804570 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" event={"ID":"6c66d4e3-dc39-4b2d-9adf-c3c69b149982","Type":"ContainerDied","Data":"ea5249ea1fd0aef7d9866a1f5dc1925a56ae07acee7d3b8930d6c12a1b621cee"} Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.805106 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea5249ea1fd0aef7d9866a1f5dc1925a56ae07acee7d3b8930d6c12a1b621cee" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.804668 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hcv8m" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.977923 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq"] Oct 02 13:26:38 crc kubenswrapper[4710]: E1002 13:26:38.978458 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="extract-content" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.978483 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="extract-content" Oct 02 13:26:38 crc kubenswrapper[4710]: E1002 13:26:38.978526 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c66d4e3-dc39-4b2d-9adf-c3c69b149982" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.978535 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c66d4e3-dc39-4b2d-9adf-c3c69b149982" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 13:26:38 crc kubenswrapper[4710]: E1002 13:26:38.978579 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="extract-utilities" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.978588 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="extract-utilities" Oct 02 13:26:38 crc kubenswrapper[4710]: E1002 13:26:38.978612 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="registry-server" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.978620 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="registry-server" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.979025 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c66d4e3-dc39-4b2d-9adf-c3c69b149982" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.979061 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5a37935-5b2e-4fb6-8c3f-8a1baa15d80c" containerName="registry-server" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.979969 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.982531 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.982606 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.982531 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.983136 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.984026 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.984836 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:26:38 crc kubenswrapper[4710]: I1002 13:26:38.988690 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq"] Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.127870 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.128078 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.128301 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.128331 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.128355 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.128579 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnk4c\" (UniqueName: \"kubernetes.io/projected/cfbcee5a-c80c-4ee3-a832-98c079200bab-kube-api-access-nnk4c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.230355 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.230414 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.230444 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.230576 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnk4c\" (UniqueName: \"kubernetes.io/projected/cfbcee5a-c80c-4ee3-a832-98c079200bab-kube-api-access-nnk4c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.230626 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.230715 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.236843 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.237184 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.237631 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.238361 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.239123 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.255598 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnk4c\" (UniqueName: \"kubernetes.io/projected/cfbcee5a-c80c-4ee3-a832-98c079200bab-kube-api-access-nnk4c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.300969 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:26:39 crc kubenswrapper[4710]: I1002 13:26:39.876895 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq"] Oct 02 13:26:40 crc kubenswrapper[4710]: I1002 13:26:40.827649 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" event={"ID":"cfbcee5a-c80c-4ee3-a832-98c079200bab","Type":"ContainerStarted","Data":"e49fcaaca5e717a3ac746154429fc2bca7c03fc0ad3ed91a8228dd5445520cec"} Oct 02 13:26:40 crc kubenswrapper[4710]: I1002 13:26:40.828296 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" event={"ID":"cfbcee5a-c80c-4ee3-a832-98c079200bab","Type":"ContainerStarted","Data":"f0eafc474f8430dc43ed857bd2c9292988a0380cbfe061d928d80febc111359f"} Oct 02 13:26:40 crc kubenswrapper[4710]: I1002 13:26:40.857663 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" podStartSLOduration=2.389031758 podStartE2EDuration="2.857633265s" podCreationTimestamp="2025-10-02 13:26:38 +0000 UTC" firstStartedPulling="2025-10-02 13:26:39.880521492 +0000 UTC m=+2243.986932375" lastFinishedPulling="2025-10-02 13:26:40.349122989 +0000 UTC m=+2244.455533882" observedRunningTime="2025-10-02 13:26:40.846991086 +0000 UTC m=+2244.953401989" watchObservedRunningTime="2025-10-02 13:26:40.857633265 +0000 UTC m=+2244.964044188" Oct 02 13:26:44 crc kubenswrapper[4710]: I1002 13:26:44.820428 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-76f964d7cf-dhr2b" podUID="407d1688-f160-4ee8-934d-9c76f353e4b9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 02 13:26:52 crc kubenswrapper[4710]: I1002 13:26:52.530100 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:26:52 crc kubenswrapper[4710]: I1002 13:26:52.530698 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:27:22 crc kubenswrapper[4710]: I1002 13:27:22.530639 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:27:22 crc kubenswrapper[4710]: I1002 13:27:22.531218 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:27:33 crc kubenswrapper[4710]: I1002 13:27:33.350488 4710 generic.go:334] "Generic (PLEG): container finished" podID="cfbcee5a-c80c-4ee3-a832-98c079200bab" containerID="e49fcaaca5e717a3ac746154429fc2bca7c03fc0ad3ed91a8228dd5445520cec" exitCode=0 Oct 02 13:27:33 crc kubenswrapper[4710]: I1002 13:27:33.350601 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" event={"ID":"cfbcee5a-c80c-4ee3-a832-98c079200bab","Type":"ContainerDied","Data":"e49fcaaca5e717a3ac746154429fc2bca7c03fc0ad3ed91a8228dd5445520cec"} Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.753803 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.869697 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-ssh-key\") pod \"cfbcee5a-c80c-4ee3-a832-98c079200bab\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.870090 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnk4c\" (UniqueName: \"kubernetes.io/projected/cfbcee5a-c80c-4ee3-a832-98c079200bab-kube-api-access-nnk4c\") pod \"cfbcee5a-c80c-4ee3-a832-98c079200bab\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.870133 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-metadata-combined-ca-bundle\") pod \"cfbcee5a-c80c-4ee3-a832-98c079200bab\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.870279 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-nova-metadata-neutron-config-0\") pod \"cfbcee5a-c80c-4ee3-a832-98c079200bab\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.870415 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-inventory\") pod \"cfbcee5a-c80c-4ee3-a832-98c079200bab\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.870480 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"cfbcee5a-c80c-4ee3-a832-98c079200bab\" (UID: \"cfbcee5a-c80c-4ee3-a832-98c079200bab\") " Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.876311 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cfbcee5a-c80c-4ee3-a832-98c079200bab" (UID: "cfbcee5a-c80c-4ee3-a832-98c079200bab"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.885959 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfbcee5a-c80c-4ee3-a832-98c079200bab-kube-api-access-nnk4c" (OuterVolumeSpecName: "kube-api-access-nnk4c") pod "cfbcee5a-c80c-4ee3-a832-98c079200bab" (UID: "cfbcee5a-c80c-4ee3-a832-98c079200bab"). InnerVolumeSpecName "kube-api-access-nnk4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.903221 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-inventory" (OuterVolumeSpecName: "inventory") pod "cfbcee5a-c80c-4ee3-a832-98c079200bab" (UID: "cfbcee5a-c80c-4ee3-a832-98c079200bab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.903275 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cfbcee5a-c80c-4ee3-a832-98c079200bab" (UID: "cfbcee5a-c80c-4ee3-a832-98c079200bab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.905907 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "cfbcee5a-c80c-4ee3-a832-98c079200bab" (UID: "cfbcee5a-c80c-4ee3-a832-98c079200bab"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.912704 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "cfbcee5a-c80c-4ee3-a832-98c079200bab" (UID: "cfbcee5a-c80c-4ee3-a832-98c079200bab"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.973512 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.973568 4710 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.973579 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.973589 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnk4c\" (UniqueName: \"kubernetes.io/projected/cfbcee5a-c80c-4ee3-a832-98c079200bab-kube-api-access-nnk4c\") on node \"crc\" DevicePath \"\"" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.973601 4710 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:27:34 crc kubenswrapper[4710]: I1002 13:27:34.973610 4710 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cfbcee5a-c80c-4ee3-a832-98c079200bab-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.370576 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" event={"ID":"cfbcee5a-c80c-4ee3-a832-98c079200bab","Type":"ContainerDied","Data":"f0eafc474f8430dc43ed857bd2c9292988a0380cbfe061d928d80febc111359f"} Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.370619 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0eafc474f8430dc43ed857bd2c9292988a0380cbfe061d928d80febc111359f" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.370638 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.466224 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787"] Oct 02 13:27:35 crc kubenswrapper[4710]: E1002 13:27:35.466700 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfbcee5a-c80c-4ee3-a832-98c079200bab" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.466723 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfbcee5a-c80c-4ee3-a832-98c079200bab" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.466955 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfbcee5a-c80c-4ee3-a832-98c079200bab" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.467814 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.470738 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.479035 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.479036 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.483231 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.483388 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.485065 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787"] Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.585714 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.585795 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45tmk\" (UniqueName: \"kubernetes.io/projected/113f767b-8234-4bc9-8ca1-03d3dd56d806-kube-api-access-45tmk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.586060 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.586182 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.586231 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.688537 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.688626 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45tmk\" (UniqueName: \"kubernetes.io/projected/113f767b-8234-4bc9-8ca1-03d3dd56d806-kube-api-access-45tmk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.688688 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.688784 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.688823 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.692837 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.693307 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.693625 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.695106 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.707788 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45tmk\" (UniqueName: \"kubernetes.io/projected/113f767b-8234-4bc9-8ca1-03d3dd56d806-kube-api-access-45tmk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q7787\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:35 crc kubenswrapper[4710]: I1002 13:27:35.804719 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:27:36 crc kubenswrapper[4710]: I1002 13:27:36.312685 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787"] Oct 02 13:27:36 crc kubenswrapper[4710]: W1002 13:27:36.318141 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod113f767b_8234_4bc9_8ca1_03d3dd56d806.slice/crio-4851d6c5f54577b8b620363f0a9a6bcb782ccd73b97727d24ab98eea766adc7a WatchSource:0}: Error finding container 4851d6c5f54577b8b620363f0a9a6bcb782ccd73b97727d24ab98eea766adc7a: Status 404 returned error can't find the container with id 4851d6c5f54577b8b620363f0a9a6bcb782ccd73b97727d24ab98eea766adc7a Oct 02 13:27:36 crc kubenswrapper[4710]: I1002 13:27:36.380420 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" event={"ID":"113f767b-8234-4bc9-8ca1-03d3dd56d806","Type":"ContainerStarted","Data":"4851d6c5f54577b8b620363f0a9a6bcb782ccd73b97727d24ab98eea766adc7a"} Oct 02 13:27:37 crc kubenswrapper[4710]: I1002 13:27:37.389721 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" event={"ID":"113f767b-8234-4bc9-8ca1-03d3dd56d806","Type":"ContainerStarted","Data":"40ff130d8af7b71d720c29a640a80df8711a76af4f79177c7553954fa9a36617"} Oct 02 13:27:37 crc kubenswrapper[4710]: I1002 13:27:37.413340 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" podStartSLOduration=1.949153892 podStartE2EDuration="2.413319256s" podCreationTimestamp="2025-10-02 13:27:35 +0000 UTC" firstStartedPulling="2025-10-02 13:27:36.320245105 +0000 UTC m=+2300.426655988" lastFinishedPulling="2025-10-02 13:27:36.784410469 +0000 UTC m=+2300.890821352" observedRunningTime="2025-10-02 13:27:37.402462572 +0000 UTC m=+2301.508873455" watchObservedRunningTime="2025-10-02 13:27:37.413319256 +0000 UTC m=+2301.519730139" Oct 02 13:27:52 crc kubenswrapper[4710]: I1002 13:27:52.530095 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:27:52 crc kubenswrapper[4710]: I1002 13:27:52.530522 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:27:52 crc kubenswrapper[4710]: I1002 13:27:52.530562 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:27:52 crc kubenswrapper[4710]: I1002 13:27:52.531291 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:27:52 crc kubenswrapper[4710]: I1002 13:27:52.531346 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" gracePeriod=600 Oct 02 13:27:52 crc kubenswrapper[4710]: E1002 13:27:52.659368 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:27:53 crc kubenswrapper[4710]: I1002 13:27:53.530756 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" exitCode=0 Oct 02 13:27:53 crc kubenswrapper[4710]: I1002 13:27:53.530895 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801"} Oct 02 13:27:53 crc kubenswrapper[4710]: I1002 13:27:53.531120 4710 scope.go:117] "RemoveContainer" containerID="e357d2ece9bd04428b3e048314016c37411ee36b3217c0372dc5abe616af8671" Oct 02 13:27:53 crc kubenswrapper[4710]: I1002 13:27:53.531455 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:27:53 crc kubenswrapper[4710]: E1002 13:27:53.531713 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:28:07 crc kubenswrapper[4710]: I1002 13:28:07.904641 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:28:07 crc kubenswrapper[4710]: E1002 13:28:07.905317 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:28:19 crc kubenswrapper[4710]: I1002 13:28:19.906728 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:28:19 crc kubenswrapper[4710]: E1002 13:28:19.907801 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:28:31 crc kubenswrapper[4710]: I1002 13:28:31.905012 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:28:31 crc kubenswrapper[4710]: E1002 13:28:31.905879 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:28:43 crc kubenswrapper[4710]: I1002 13:28:43.905079 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:28:43 crc kubenswrapper[4710]: E1002 13:28:43.905894 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:28:58 crc kubenswrapper[4710]: I1002 13:28:58.904375 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:28:58 crc kubenswrapper[4710]: E1002 13:28:58.905343 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:29:09 crc kubenswrapper[4710]: I1002 13:29:09.905596 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:29:09 crc kubenswrapper[4710]: E1002 13:29:09.906704 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:29:20 crc kubenswrapper[4710]: I1002 13:29:20.905188 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:29:20 crc kubenswrapper[4710]: E1002 13:29:20.905946 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:29:35 crc kubenswrapper[4710]: I1002 13:29:35.905183 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:29:35 crc kubenswrapper[4710]: E1002 13:29:35.906160 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:29:49 crc kubenswrapper[4710]: I1002 13:29:49.904486 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:29:49 crc kubenswrapper[4710]: E1002 13:29:49.905358 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.156312 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72"] Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.158416 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.163408 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.163805 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.175391 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72"] Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.287428 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-secret-volume\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.287997 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcljq\" (UniqueName: \"kubernetes.io/projected/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-kube-api-access-pcljq\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.288235 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-config-volume\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.390360 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcljq\" (UniqueName: \"kubernetes.io/projected/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-kube-api-access-pcljq\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.390554 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-config-volume\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.390657 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-secret-volume\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.391941 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-config-volume\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.399701 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-secret-volume\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.411670 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcljq\" (UniqueName: \"kubernetes.io/projected/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-kube-api-access-pcljq\") pod \"collect-profiles-29323530-8zt72\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.479192 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:00 crc kubenswrapper[4710]: I1002 13:30:00.932421 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72"] Oct 02 13:30:01 crc kubenswrapper[4710]: I1002 13:30:01.835395 4710 generic.go:334] "Generic (PLEG): container finished" podID="e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" containerID="4572e5a2057d55eabe5df7321a3aba9b8e05d18c81c9682c5c81fb53f28cad65" exitCode=0 Oct 02 13:30:01 crc kubenswrapper[4710]: I1002 13:30:01.835691 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" event={"ID":"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f","Type":"ContainerDied","Data":"4572e5a2057d55eabe5df7321a3aba9b8e05d18c81c9682c5c81fb53f28cad65"} Oct 02 13:30:01 crc kubenswrapper[4710]: I1002 13:30:01.835720 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" event={"ID":"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f","Type":"ContainerStarted","Data":"0a8a8d2112f5f82ed20ad42fa88376382711bc662f64158eccd9c459f4327c77"} Oct 02 13:30:02 crc kubenswrapper[4710]: I1002 13:30:02.905102 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:30:02 crc kubenswrapper[4710]: E1002 13:30:02.906306 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.168828 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.254389 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-secret-volume\") pod \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.254556 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcljq\" (UniqueName: \"kubernetes.io/projected/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-kube-api-access-pcljq\") pod \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.254796 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-config-volume\") pod \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\" (UID: \"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f\") " Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.255414 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-config-volume" (OuterVolumeSpecName: "config-volume") pod "e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" (UID: "e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.255978 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.260538 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-kube-api-access-pcljq" (OuterVolumeSpecName: "kube-api-access-pcljq") pod "e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" (UID: "e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f"). InnerVolumeSpecName "kube-api-access-pcljq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.260861 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" (UID: "e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.359393 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.359454 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcljq\" (UniqueName: \"kubernetes.io/projected/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f-kube-api-access-pcljq\") on node \"crc\" DevicePath \"\"" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.859159 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" event={"ID":"e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f","Type":"ContainerDied","Data":"0a8a8d2112f5f82ed20ad42fa88376382711bc662f64158eccd9c459f4327c77"} Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.859209 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a8a8d2112f5f82ed20ad42fa88376382711bc662f64158eccd9c459f4327c77" Oct 02 13:30:03 crc kubenswrapper[4710]: I1002 13:30:03.859245 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72" Oct 02 13:30:04 crc kubenswrapper[4710]: I1002 13:30:04.239018 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz"] Oct 02 13:30:04 crc kubenswrapper[4710]: I1002 13:30:04.247044 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323485-9pxkz"] Oct 02 13:30:04 crc kubenswrapper[4710]: I1002 13:30:04.939250 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8205f2e-3067-4ab0-9e30-43d99dfd53eb" path="/var/lib/kubelet/pods/e8205f2e-3067-4ab0-9e30-43d99dfd53eb/volumes" Oct 02 13:30:15 crc kubenswrapper[4710]: I1002 13:30:15.904704 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:30:15 crc kubenswrapper[4710]: E1002 13:30:15.905340 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:30:29 crc kubenswrapper[4710]: I1002 13:30:29.905412 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:30:29 crc kubenswrapper[4710]: E1002 13:30:29.906612 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:30:42 crc kubenswrapper[4710]: I1002 13:30:42.905721 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:30:42 crc kubenswrapper[4710]: E1002 13:30:42.907568 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:30:47 crc kubenswrapper[4710]: I1002 13:30:47.284022 4710 scope.go:117] "RemoveContainer" containerID="615d81488b245ad7e8fb43787023141b4ba3fc1d197076bceafe1fe7a4195f68" Oct 02 13:30:55 crc kubenswrapper[4710]: I1002 13:30:55.904272 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:30:55 crc kubenswrapper[4710]: E1002 13:30:55.904993 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:31:09 crc kubenswrapper[4710]: I1002 13:31:09.904962 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:31:09 crc kubenswrapper[4710]: E1002 13:31:09.905640 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:31:24 crc kubenswrapper[4710]: I1002 13:31:24.904823 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:31:24 crc kubenswrapper[4710]: E1002 13:31:24.905939 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:31:35 crc kubenswrapper[4710]: I1002 13:31:35.905619 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:31:35 crc kubenswrapper[4710]: E1002 13:31:35.906614 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:31:48 crc kubenswrapper[4710]: I1002 13:31:48.905257 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:31:48 crc kubenswrapper[4710]: E1002 13:31:48.906945 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:31:59 crc kubenswrapper[4710]: I1002 13:31:59.904737 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:31:59 crc kubenswrapper[4710]: E1002 13:31:59.905872 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:32:05 crc kubenswrapper[4710]: I1002 13:32:05.997138 4710 generic.go:334] "Generic (PLEG): container finished" podID="113f767b-8234-4bc9-8ca1-03d3dd56d806" containerID="40ff130d8af7b71d720c29a640a80df8711a76af4f79177c7553954fa9a36617" exitCode=0 Oct 02 13:32:05 crc kubenswrapper[4710]: I1002 13:32:05.997189 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" event={"ID":"113f767b-8234-4bc9-8ca1-03d3dd56d806","Type":"ContainerDied","Data":"40ff130d8af7b71d720c29a640a80df8711a76af4f79177c7553954fa9a36617"} Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.429954 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.574686 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-ssh-key\") pod \"113f767b-8234-4bc9-8ca1-03d3dd56d806\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.574784 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45tmk\" (UniqueName: \"kubernetes.io/projected/113f767b-8234-4bc9-8ca1-03d3dd56d806-kube-api-access-45tmk\") pod \"113f767b-8234-4bc9-8ca1-03d3dd56d806\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.574879 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-inventory\") pod \"113f767b-8234-4bc9-8ca1-03d3dd56d806\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.574989 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-secret-0\") pod \"113f767b-8234-4bc9-8ca1-03d3dd56d806\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.575018 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-combined-ca-bundle\") pod \"113f767b-8234-4bc9-8ca1-03d3dd56d806\" (UID: \"113f767b-8234-4bc9-8ca1-03d3dd56d806\") " Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.580303 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/113f767b-8234-4bc9-8ca1-03d3dd56d806-kube-api-access-45tmk" (OuterVolumeSpecName: "kube-api-access-45tmk") pod "113f767b-8234-4bc9-8ca1-03d3dd56d806" (UID: "113f767b-8234-4bc9-8ca1-03d3dd56d806"). InnerVolumeSpecName "kube-api-access-45tmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.581653 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "113f767b-8234-4bc9-8ca1-03d3dd56d806" (UID: "113f767b-8234-4bc9-8ca1-03d3dd56d806"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.609481 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-inventory" (OuterVolumeSpecName: "inventory") pod "113f767b-8234-4bc9-8ca1-03d3dd56d806" (UID: "113f767b-8234-4bc9-8ca1-03d3dd56d806"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.614761 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "113f767b-8234-4bc9-8ca1-03d3dd56d806" (UID: "113f767b-8234-4bc9-8ca1-03d3dd56d806"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.616136 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "113f767b-8234-4bc9-8ca1-03d3dd56d806" (UID: "113f767b-8234-4bc9-8ca1-03d3dd56d806"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.680609 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.680652 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45tmk\" (UniqueName: \"kubernetes.io/projected/113f767b-8234-4bc9-8ca1-03d3dd56d806-kube-api-access-45tmk\") on node \"crc\" DevicePath \"\"" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.680667 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.680680 4710 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:32:07 crc kubenswrapper[4710]: I1002 13:32:07.680695 4710 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/113f767b-8234-4bc9-8ca1-03d3dd56d806-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.013965 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" event={"ID":"113f767b-8234-4bc9-8ca1-03d3dd56d806","Type":"ContainerDied","Data":"4851d6c5f54577b8b620363f0a9a6bcb782ccd73b97727d24ab98eea766adc7a"} Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.014012 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4851d6c5f54577b8b620363f0a9a6bcb782ccd73b97727d24ab98eea766adc7a" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.014050 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q7787" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.120818 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d"] Oct 02 13:32:08 crc kubenswrapper[4710]: E1002 13:32:08.121783 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113f767b-8234-4bc9-8ca1-03d3dd56d806" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.121804 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="113f767b-8234-4bc9-8ca1-03d3dd56d806" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 13:32:08 crc kubenswrapper[4710]: E1002 13:32:08.121857 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" containerName="collect-profiles" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.121870 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" containerName="collect-profiles" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.122136 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="113f767b-8234-4bc9-8ca1-03d3dd56d806" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.122157 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" containerName="collect-profiles" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.122882 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.124646 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.124673 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.126140 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.126441 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.126657 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.126847 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.130077 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.151003 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d"] Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.290993 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291081 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291115 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291135 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291169 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tk6t\" (UniqueName: \"kubernetes.io/projected/ab353848-8250-4624-adeb-72e97fdfdb30-kube-api-access-9tk6t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291203 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291224 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ab353848-8250-4624-adeb-72e97fdfdb30-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291246 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.291275 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.392939 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393024 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393058 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393077 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393106 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tk6t\" (UniqueName: \"kubernetes.io/projected/ab353848-8250-4624-adeb-72e97fdfdb30-kube-api-access-9tk6t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393135 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393156 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ab353848-8250-4624-adeb-72e97fdfdb30-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393176 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.393223 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.394975 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ab353848-8250-4624-adeb-72e97fdfdb30-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.397233 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.398009 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.398054 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.398193 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.398667 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.399082 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.407319 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.410631 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tk6t\" (UniqueName: \"kubernetes.io/projected/ab353848-8250-4624-adeb-72e97fdfdb30-kube-api-access-9tk6t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-brx4d\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.451424 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:32:08 crc kubenswrapper[4710]: I1002 13:32:08.995666 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d"] Oct 02 13:32:09 crc kubenswrapper[4710]: I1002 13:32:09.002602 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:32:09 crc kubenswrapper[4710]: I1002 13:32:09.027267 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" event={"ID":"ab353848-8250-4624-adeb-72e97fdfdb30","Type":"ContainerStarted","Data":"5e264bb57edce6207e5a2cf1510a4bfb48077687d545b1bf56b2e504aebe6382"} Oct 02 13:32:10 crc kubenswrapper[4710]: I1002 13:32:10.037516 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" event={"ID":"ab353848-8250-4624-adeb-72e97fdfdb30","Type":"ContainerStarted","Data":"55035cc4eeb766a04fbde9746ea78112519cf5e765b45f8d88ac7266ca5e34ed"} Oct 02 13:32:10 crc kubenswrapper[4710]: I1002 13:32:10.062610 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" podStartSLOduration=1.5950274869999999 podStartE2EDuration="2.062593529s" podCreationTimestamp="2025-10-02 13:32:08 +0000 UTC" firstStartedPulling="2025-10-02 13:32:09.002333679 +0000 UTC m=+2573.108744562" lastFinishedPulling="2025-10-02 13:32:09.469899721 +0000 UTC m=+2573.576310604" observedRunningTime="2025-10-02 13:32:10.062009534 +0000 UTC m=+2574.168420437" watchObservedRunningTime="2025-10-02 13:32:10.062593529 +0000 UTC m=+2574.169004412" Oct 02 13:32:10 crc kubenswrapper[4710]: I1002 13:32:10.904730 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:32:10 crc kubenswrapper[4710]: E1002 13:32:10.905627 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:32:24 crc kubenswrapper[4710]: I1002 13:32:24.905179 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:32:24 crc kubenswrapper[4710]: E1002 13:32:24.906410 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:32:35 crc kubenswrapper[4710]: I1002 13:32:35.905090 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:32:35 crc kubenswrapper[4710]: E1002 13:32:35.905805 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:32:48 crc kubenswrapper[4710]: I1002 13:32:48.905169 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:32:48 crc kubenswrapper[4710]: E1002 13:32:48.906301 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:33:03 crc kubenswrapper[4710]: I1002 13:33:03.905159 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:33:04 crc kubenswrapper[4710]: I1002 13:33:04.557241 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"401a1c1daeea58980316aac659a5f50b88f33db73a751be469657a6bb2849ad4"} Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.099759 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8zjz6"] Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.103533 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.108133 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zjz6"] Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.223644 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-catalog-content\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.223729 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-utilities\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.224531 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6mrd\" (UniqueName: \"kubernetes.io/projected/c565d718-de82-4e56-b0ae-8c0f3432eab9-kube-api-access-k6mrd\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.326686 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-catalog-content\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.326815 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-utilities\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.327010 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6mrd\" (UniqueName: \"kubernetes.io/projected/c565d718-de82-4e56-b0ae-8c0f3432eab9-kube-api-access-k6mrd\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.327220 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-catalog-content\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.327476 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-utilities\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.355989 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6mrd\" (UniqueName: \"kubernetes.io/projected/c565d718-de82-4e56-b0ae-8c0f3432eab9-kube-api-access-k6mrd\") pod \"redhat-operators-8zjz6\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.432451 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:28 crc kubenswrapper[4710]: I1002 13:34:28.949553 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zjz6"] Oct 02 13:34:29 crc kubenswrapper[4710]: I1002 13:34:29.377801 4710 generic.go:334] "Generic (PLEG): container finished" podID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerID="0f8df94fdd0bab2750e43883fdc5fa52745ee1f19c44b5c1f5a6a42c88f52a47" exitCode=0 Oct 02 13:34:29 crc kubenswrapper[4710]: I1002 13:34:29.377859 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerDied","Data":"0f8df94fdd0bab2750e43883fdc5fa52745ee1f19c44b5c1f5a6a42c88f52a47"} Oct 02 13:34:29 crc kubenswrapper[4710]: I1002 13:34:29.378078 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerStarted","Data":"9f0500a3f0e93aaa17108811d5400f66208b7f27e4e48829693b85983e92ec5f"} Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.685897 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-96w82"] Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.689286 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.702170 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96w82"] Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.778676 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-catalog-content\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.778771 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-utilities\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.779459 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h95fr\" (UniqueName: \"kubernetes.io/projected/47f7aa13-e280-4cae-be08-8a5ee396aa0d-kube-api-access-h95fr\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.881666 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h95fr\" (UniqueName: \"kubernetes.io/projected/47f7aa13-e280-4cae-be08-8a5ee396aa0d-kube-api-access-h95fr\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.882063 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-catalog-content\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.882097 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-utilities\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.882584 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-catalog-content\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.882630 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-utilities\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:30 crc kubenswrapper[4710]: I1002 13:34:30.912212 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h95fr\" (UniqueName: \"kubernetes.io/projected/47f7aa13-e280-4cae-be08-8a5ee396aa0d-kube-api-access-h95fr\") pod \"redhat-marketplace-96w82\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:31 crc kubenswrapper[4710]: I1002 13:34:31.083384 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:31 crc kubenswrapper[4710]: I1002 13:34:31.398684 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerStarted","Data":"8d7c8dffc64395f7900945f35aedaef814534d8f34ccf7e457d694f55fee5e12"} Oct 02 13:34:31 crc kubenswrapper[4710]: I1002 13:34:31.554204 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96w82"] Oct 02 13:34:31 crc kubenswrapper[4710]: W1002 13:34:31.556440 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47f7aa13_e280_4cae_be08_8a5ee396aa0d.slice/crio-7ef8bf39dfaf60cffcd6fda6d6b854eb1369c7751a742a28a3a1fc47631508b6 WatchSource:0}: Error finding container 7ef8bf39dfaf60cffcd6fda6d6b854eb1369c7751a742a28a3a1fc47631508b6: Status 404 returned error can't find the container with id 7ef8bf39dfaf60cffcd6fda6d6b854eb1369c7751a742a28a3a1fc47631508b6 Oct 02 13:34:32 crc kubenswrapper[4710]: I1002 13:34:32.418165 4710 generic.go:334] "Generic (PLEG): container finished" podID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerID="8d7c8dffc64395f7900945f35aedaef814534d8f34ccf7e457d694f55fee5e12" exitCode=0 Oct 02 13:34:32 crc kubenswrapper[4710]: I1002 13:34:32.418805 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerDied","Data":"8d7c8dffc64395f7900945f35aedaef814534d8f34ccf7e457d694f55fee5e12"} Oct 02 13:34:32 crc kubenswrapper[4710]: I1002 13:34:32.426367 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96w82" event={"ID":"47f7aa13-e280-4cae-be08-8a5ee396aa0d","Type":"ContainerStarted","Data":"7ef8bf39dfaf60cffcd6fda6d6b854eb1369c7751a742a28a3a1fc47631508b6"} Oct 02 13:34:33 crc kubenswrapper[4710]: I1002 13:34:33.451869 4710 generic.go:334] "Generic (PLEG): container finished" podID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerID="02c64043dd7791e2877c1a2340d01e2c3d4639793891a06252b50b22bad05588" exitCode=0 Oct 02 13:34:33 crc kubenswrapper[4710]: I1002 13:34:33.452223 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96w82" event={"ID":"47f7aa13-e280-4cae-be08-8a5ee396aa0d","Type":"ContainerDied","Data":"02c64043dd7791e2877c1a2340d01e2c3d4639793891a06252b50b22bad05588"} Oct 02 13:34:36 crc kubenswrapper[4710]: I1002 13:34:36.482892 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerStarted","Data":"5bfd229ce406265397980526857f49935afef9177a1d6bf1b2bbaf2773df175a"} Oct 02 13:34:37 crc kubenswrapper[4710]: I1002 13:34:37.521894 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8zjz6" podStartSLOduration=2.943508312 podStartE2EDuration="9.521873676s" podCreationTimestamp="2025-10-02 13:34:28 +0000 UTC" firstStartedPulling="2025-10-02 13:34:29.379592817 +0000 UTC m=+2713.486003700" lastFinishedPulling="2025-10-02 13:34:35.957958161 +0000 UTC m=+2720.064369064" observedRunningTime="2025-10-02 13:34:37.515767271 +0000 UTC m=+2721.622178174" watchObservedRunningTime="2025-10-02 13:34:37.521873676 +0000 UTC m=+2721.628284559" Oct 02 13:34:38 crc kubenswrapper[4710]: I1002 13:34:38.434034 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:38 crc kubenswrapper[4710]: I1002 13:34:38.434559 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:38 crc kubenswrapper[4710]: I1002 13:34:38.501787 4710 generic.go:334] "Generic (PLEG): container finished" podID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerID="2c70cc29ffe895201925454d271645ee9082e769c35ae836a2432ef5b240a8cb" exitCode=0 Oct 02 13:34:38 crc kubenswrapper[4710]: I1002 13:34:38.502871 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96w82" event={"ID":"47f7aa13-e280-4cae-be08-8a5ee396aa0d","Type":"ContainerDied","Data":"2c70cc29ffe895201925454d271645ee9082e769c35ae836a2432ef5b240a8cb"} Oct 02 13:34:39 crc kubenswrapper[4710]: I1002 13:34:39.486924 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8zjz6" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="registry-server" probeResult="failure" output=< Oct 02 13:34:39 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 13:34:39 crc kubenswrapper[4710]: > Oct 02 13:34:41 crc kubenswrapper[4710]: I1002 13:34:41.535938 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96w82" event={"ID":"47f7aa13-e280-4cae-be08-8a5ee396aa0d","Type":"ContainerStarted","Data":"89ccb492130b67f244cf132665d90be94c61da388925dc7d88417a2126debc57"} Oct 02 13:34:41 crc kubenswrapper[4710]: I1002 13:34:41.563688 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-96w82" podStartSLOduration=4.727558048 podStartE2EDuration="11.563664673s" podCreationTimestamp="2025-10-02 13:34:30 +0000 UTC" firstStartedPulling="2025-10-02 13:34:33.703362169 +0000 UTC m=+2717.809773052" lastFinishedPulling="2025-10-02 13:34:40.539468794 +0000 UTC m=+2724.645879677" observedRunningTime="2025-10-02 13:34:41.551663799 +0000 UTC m=+2725.658074682" watchObservedRunningTime="2025-10-02 13:34:41.563664673 +0000 UTC m=+2725.670075556" Oct 02 13:34:48 crc kubenswrapper[4710]: I1002 13:34:48.479389 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:48 crc kubenswrapper[4710]: I1002 13:34:48.529730 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:48 crc kubenswrapper[4710]: I1002 13:34:48.720771 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zjz6"] Oct 02 13:34:49 crc kubenswrapper[4710]: I1002 13:34:49.606206 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8zjz6" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="registry-server" containerID="cri-o://5bfd229ce406265397980526857f49935afef9177a1d6bf1b2bbaf2773df175a" gracePeriod=2 Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.617829 4710 generic.go:334] "Generic (PLEG): container finished" podID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerID="5bfd229ce406265397980526857f49935afef9177a1d6bf1b2bbaf2773df175a" exitCode=0 Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.618003 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerDied","Data":"5bfd229ce406265397980526857f49935afef9177a1d6bf1b2bbaf2773df175a"} Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.783515 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.805852 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6mrd\" (UniqueName: \"kubernetes.io/projected/c565d718-de82-4e56-b0ae-8c0f3432eab9-kube-api-access-k6mrd\") pod \"c565d718-de82-4e56-b0ae-8c0f3432eab9\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.806045 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-catalog-content\") pod \"c565d718-de82-4e56-b0ae-8c0f3432eab9\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.806090 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-utilities\") pod \"c565d718-de82-4e56-b0ae-8c0f3432eab9\" (UID: \"c565d718-de82-4e56-b0ae-8c0f3432eab9\") " Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.807199 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-utilities" (OuterVolumeSpecName: "utilities") pod "c565d718-de82-4e56-b0ae-8c0f3432eab9" (UID: "c565d718-de82-4e56-b0ae-8c0f3432eab9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.811219 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c565d718-de82-4e56-b0ae-8c0f3432eab9-kube-api-access-k6mrd" (OuterVolumeSpecName: "kube-api-access-k6mrd") pod "c565d718-de82-4e56-b0ae-8c0f3432eab9" (UID: "c565d718-de82-4e56-b0ae-8c0f3432eab9"). InnerVolumeSpecName "kube-api-access-k6mrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.883825 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c565d718-de82-4e56-b0ae-8c0f3432eab9" (UID: "c565d718-de82-4e56-b0ae-8c0f3432eab9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.908207 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.908236 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c565d718-de82-4e56-b0ae-8c0f3432eab9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:34:50 crc kubenswrapper[4710]: I1002 13:34:50.908248 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6mrd\" (UniqueName: \"kubernetes.io/projected/c565d718-de82-4e56-b0ae-8c0f3432eab9-kube-api-access-k6mrd\") on node \"crc\" DevicePath \"\"" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.083972 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.084052 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.138808 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.630771 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjz6" event={"ID":"c565d718-de82-4e56-b0ae-8c0f3432eab9","Type":"ContainerDied","Data":"9f0500a3f0e93aaa17108811d5400f66208b7f27e4e48829693b85983e92ec5f"} Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.631169 4710 scope.go:117] "RemoveContainer" containerID="5bfd229ce406265397980526857f49935afef9177a1d6bf1b2bbaf2773df175a" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.630796 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjz6" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.665163 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zjz6"] Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.669283 4710 scope.go:117] "RemoveContainer" containerID="8d7c8dffc64395f7900945f35aedaef814534d8f34ccf7e457d694f55fee5e12" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.673149 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8zjz6"] Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.696845 4710 scope.go:117] "RemoveContainer" containerID="0f8df94fdd0bab2750e43883fdc5fa52745ee1f19c44b5c1f5a6a42c88f52a47" Oct 02 13:34:51 crc kubenswrapper[4710]: I1002 13:34:51.700277 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:52 crc kubenswrapper[4710]: I1002 13:34:52.513573 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96w82"] Oct 02 13:34:52 crc kubenswrapper[4710]: I1002 13:34:52.916988 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" path="/var/lib/kubelet/pods/c565d718-de82-4e56-b0ae-8c0f3432eab9/volumes" Oct 02 13:34:53 crc kubenswrapper[4710]: I1002 13:34:53.649627 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-96w82" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="registry-server" containerID="cri-o://89ccb492130b67f244cf132665d90be94c61da388925dc7d88417a2126debc57" gracePeriod=2 Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.663022 4710 generic.go:334] "Generic (PLEG): container finished" podID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerID="89ccb492130b67f244cf132665d90be94c61da388925dc7d88417a2126debc57" exitCode=0 Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.663083 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96w82" event={"ID":"47f7aa13-e280-4cae-be08-8a5ee396aa0d","Type":"ContainerDied","Data":"89ccb492130b67f244cf132665d90be94c61da388925dc7d88417a2126debc57"} Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.663661 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96w82" event={"ID":"47f7aa13-e280-4cae-be08-8a5ee396aa0d","Type":"ContainerDied","Data":"7ef8bf39dfaf60cffcd6fda6d6b854eb1369c7751a742a28a3a1fc47631508b6"} Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.663681 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ef8bf39dfaf60cffcd6fda6d6b854eb1369c7751a742a28a3a1fc47631508b6" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.702109 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.784870 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-catalog-content\") pod \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.785154 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h95fr\" (UniqueName: \"kubernetes.io/projected/47f7aa13-e280-4cae-be08-8a5ee396aa0d-kube-api-access-h95fr\") pod \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.785297 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-utilities\") pod \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\" (UID: \"47f7aa13-e280-4cae-be08-8a5ee396aa0d\") " Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.786427 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-utilities" (OuterVolumeSpecName: "utilities") pod "47f7aa13-e280-4cae-be08-8a5ee396aa0d" (UID: "47f7aa13-e280-4cae-be08-8a5ee396aa0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.805086 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47f7aa13-e280-4cae-be08-8a5ee396aa0d-kube-api-access-h95fr" (OuterVolumeSpecName: "kube-api-access-h95fr") pod "47f7aa13-e280-4cae-be08-8a5ee396aa0d" (UID: "47f7aa13-e280-4cae-be08-8a5ee396aa0d"). InnerVolumeSpecName "kube-api-access-h95fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.815957 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47f7aa13-e280-4cae-be08-8a5ee396aa0d" (UID: "47f7aa13-e280-4cae-be08-8a5ee396aa0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.889136 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.889173 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h95fr\" (UniqueName: \"kubernetes.io/projected/47f7aa13-e280-4cae-be08-8a5ee396aa0d-kube-api-access-h95fr\") on node \"crc\" DevicePath \"\"" Oct 02 13:34:54 crc kubenswrapper[4710]: I1002 13:34:54.889185 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f7aa13-e280-4cae-be08-8a5ee396aa0d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:34:55 crc kubenswrapper[4710]: I1002 13:34:55.680513 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96w82" Oct 02 13:34:55 crc kubenswrapper[4710]: I1002 13:34:55.708762 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96w82"] Oct 02 13:34:55 crc kubenswrapper[4710]: I1002 13:34:55.717888 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-96w82"] Oct 02 13:34:56 crc kubenswrapper[4710]: I1002 13:34:56.919126 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" path="/var/lib/kubelet/pods/47f7aa13-e280-4cae-be08-8a5ee396aa0d/volumes" Oct 02 13:35:22 crc kubenswrapper[4710]: I1002 13:35:22.530732 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:35:22 crc kubenswrapper[4710]: I1002 13:35:22.532286 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:35:52 crc kubenswrapper[4710]: I1002 13:35:52.530906 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:35:52 crc kubenswrapper[4710]: I1002 13:35:52.531738 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:35:55 crc kubenswrapper[4710]: I1002 13:35:55.277450 4710 generic.go:334] "Generic (PLEG): container finished" podID="ab353848-8250-4624-adeb-72e97fdfdb30" containerID="55035cc4eeb766a04fbde9746ea78112519cf5e765b45f8d88ac7266ca5e34ed" exitCode=0 Oct 02 13:35:55 crc kubenswrapper[4710]: I1002 13:35:55.277696 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" event={"ID":"ab353848-8250-4624-adeb-72e97fdfdb30","Type":"ContainerDied","Data":"55035cc4eeb766a04fbde9746ea78112519cf5e765b45f8d88ac7266ca5e34ed"} Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.699328 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.796125 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-0\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.797943 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tk6t\" (UniqueName: \"kubernetes.io/projected/ab353848-8250-4624-adeb-72e97fdfdb30-kube-api-access-9tk6t\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798073 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-combined-ca-bundle\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798116 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-ssh-key\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798141 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-1\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798196 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-1\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798271 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ab353848-8250-4624-adeb-72e97fdfdb30-nova-extra-config-0\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798296 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-inventory\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.798326 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-0\") pod \"ab353848-8250-4624-adeb-72e97fdfdb30\" (UID: \"ab353848-8250-4624-adeb-72e97fdfdb30\") " Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.804147 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.805849 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab353848-8250-4624-adeb-72e97fdfdb30-kube-api-access-9tk6t" (OuterVolumeSpecName: "kube-api-access-9tk6t") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "kube-api-access-9tk6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.830667 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.830795 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.833160 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.836008 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-inventory" (OuterVolumeSpecName: "inventory") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.836804 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab353848-8250-4624-adeb-72e97fdfdb30-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.840059 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.841082 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab353848-8250-4624-adeb-72e97fdfdb30" (UID: "ab353848-8250-4624-adeb-72e97fdfdb30"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901269 4710 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901388 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tk6t\" (UniqueName: \"kubernetes.io/projected/ab353848-8250-4624-adeb-72e97fdfdb30-kube-api-access-9tk6t\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901407 4710 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901421 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901432 4710 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901444 4710 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901457 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901469 4710 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ab353848-8250-4624-adeb-72e97fdfdb30-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:56 crc kubenswrapper[4710]: I1002 13:35:56.901479 4710 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ab353848-8250-4624-adeb-72e97fdfdb30-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.301239 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" event={"ID":"ab353848-8250-4624-adeb-72e97fdfdb30","Type":"ContainerDied","Data":"5e264bb57edce6207e5a2cf1510a4bfb48077687d545b1bf56b2e504aebe6382"} Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.301302 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e264bb57edce6207e5a2cf1510a4bfb48077687d545b1bf56b2e504aebe6382" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.301384 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-brx4d" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.419893 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h"] Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420303 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="extract-utilities" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420323 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="extract-utilities" Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420335 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="extract-utilities" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420342 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="extract-utilities" Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420359 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="registry-server" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420366 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="registry-server" Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420380 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab353848-8250-4624-adeb-72e97fdfdb30" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420386 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab353848-8250-4624-adeb-72e97fdfdb30" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420402 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="registry-server" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420421 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="registry-server" Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420445 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="extract-content" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420451 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="extract-content" Oct 02 13:35:57 crc kubenswrapper[4710]: E1002 13:35:57.420460 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="extract-content" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420466 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="extract-content" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420645 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c565d718-de82-4e56-b0ae-8c0f3432eab9" containerName="registry-server" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420664 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f7aa13-e280-4cae-be08-8a5ee396aa0d" containerName="registry-server" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.420674 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab353848-8250-4624-adeb-72e97fdfdb30" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.421338 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.423456 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.432228 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.432358 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.432384 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qgzcw" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.437629 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.438151 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h"] Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.517579 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.517648 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.517726 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnwlc\" (UniqueName: \"kubernetes.io/projected/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-kube-api-access-jnwlc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.518227 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.518449 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.518533 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.518631 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621147 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621220 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621279 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnwlc\" (UniqueName: \"kubernetes.io/projected/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-kube-api-access-jnwlc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621357 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621387 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621409 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.621437 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.625460 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.625807 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.625973 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.626673 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.629566 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.633626 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.649654 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnwlc\" (UniqueName: \"kubernetes.io/projected/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-kube-api-access-jnwlc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:57 crc kubenswrapper[4710]: I1002 13:35:57.740486 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:35:58 crc kubenswrapper[4710]: I1002 13:35:58.315611 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h"] Oct 02 13:35:59 crc kubenswrapper[4710]: I1002 13:35:59.324776 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" event={"ID":"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8","Type":"ContainerStarted","Data":"793f842deb770c4035ba8c3fd5d96b32608969a94124a574915d77b78f39953d"} Oct 02 13:35:59 crc kubenswrapper[4710]: I1002 13:35:59.325132 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" event={"ID":"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8","Type":"ContainerStarted","Data":"48a9faa8fca8a4c694569c28293d69933efc744702602cc5462e0a3a21611bb6"} Oct 02 13:35:59 crc kubenswrapper[4710]: I1002 13:35:59.354217 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" podStartSLOduration=1.878011865 podStartE2EDuration="2.354191945s" podCreationTimestamp="2025-10-02 13:35:57 +0000 UTC" firstStartedPulling="2025-10-02 13:35:58.333244499 +0000 UTC m=+2802.439655382" lastFinishedPulling="2025-10-02 13:35:58.809424539 +0000 UTC m=+2802.915835462" observedRunningTime="2025-10-02 13:35:59.35004659 +0000 UTC m=+2803.456457463" watchObservedRunningTime="2025-10-02 13:35:59.354191945 +0000 UTC m=+2803.460602848" Oct 02 13:36:09 crc kubenswrapper[4710]: I1002 13:36:09.895207 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-75pwg"] Oct 02 13:36:09 crc kubenswrapper[4710]: I1002 13:36:09.898558 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:09 crc kubenswrapper[4710]: I1002 13:36:09.915104 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-75pwg"] Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.088275 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-catalog-content\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.088799 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-utilities\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.089426 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfvxs\" (UniqueName: \"kubernetes.io/projected/a9cb4718-1ed0-40d9-8378-b3221ce3d529-kube-api-access-kfvxs\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.192870 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfvxs\" (UniqueName: \"kubernetes.io/projected/a9cb4718-1ed0-40d9-8378-b3221ce3d529-kube-api-access-kfvxs\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.193478 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-catalog-content\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.194114 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-catalog-content\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.194451 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-utilities\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.194493 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-utilities\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.218295 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfvxs\" (UniqueName: \"kubernetes.io/projected/a9cb4718-1ed0-40d9-8378-b3221ce3d529-kube-api-access-kfvxs\") pod \"certified-operators-75pwg\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.223507 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:10 crc kubenswrapper[4710]: I1002 13:36:10.770497 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-75pwg"] Oct 02 13:36:11 crc kubenswrapper[4710]: I1002 13:36:11.454331 4710 generic.go:334] "Generic (PLEG): container finished" podID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerID="82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111" exitCode=0 Oct 02 13:36:11 crc kubenswrapper[4710]: I1002 13:36:11.454474 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerDied","Data":"82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111"} Oct 02 13:36:11 crc kubenswrapper[4710]: I1002 13:36:11.454642 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerStarted","Data":"caae8e517b7caba82a4f17c5c02cc1041892db938bf4d35d2a791bc38e87ae5c"} Oct 02 13:36:13 crc kubenswrapper[4710]: I1002 13:36:13.489504 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerStarted","Data":"01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05"} Oct 02 13:36:14 crc kubenswrapper[4710]: I1002 13:36:14.499586 4710 generic.go:334] "Generic (PLEG): container finished" podID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerID="01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05" exitCode=0 Oct 02 13:36:14 crc kubenswrapper[4710]: I1002 13:36:14.499627 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerDied","Data":"01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05"} Oct 02 13:36:15 crc kubenswrapper[4710]: I1002 13:36:15.510041 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerStarted","Data":"5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f"} Oct 02 13:36:15 crc kubenswrapper[4710]: I1002 13:36:15.533074 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-75pwg" podStartSLOduration=3.020139004 podStartE2EDuration="6.533056988s" podCreationTimestamp="2025-10-02 13:36:09 +0000 UTC" firstStartedPulling="2025-10-02 13:36:11.456718168 +0000 UTC m=+2815.563129051" lastFinishedPulling="2025-10-02 13:36:14.969636142 +0000 UTC m=+2819.076047035" observedRunningTime="2025-10-02 13:36:15.532391252 +0000 UTC m=+2819.638802125" watchObservedRunningTime="2025-10-02 13:36:15.533056988 +0000 UTC m=+2819.639467871" Oct 02 13:36:20 crc kubenswrapper[4710]: I1002 13:36:20.224424 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:20 crc kubenswrapper[4710]: I1002 13:36:20.225019 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:20 crc kubenswrapper[4710]: I1002 13:36:20.272375 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:20 crc kubenswrapper[4710]: I1002 13:36:20.611239 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:20 crc kubenswrapper[4710]: I1002 13:36:20.661271 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-75pwg"] Oct 02 13:36:22 crc kubenswrapper[4710]: I1002 13:36:22.531008 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:36:22 crc kubenswrapper[4710]: I1002 13:36:22.531302 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:36:22 crc kubenswrapper[4710]: I1002 13:36:22.531356 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:36:22 crc kubenswrapper[4710]: I1002 13:36:22.532249 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"401a1c1daeea58980316aac659a5f50b88f33db73a751be469657a6bb2849ad4"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:36:22 crc kubenswrapper[4710]: I1002 13:36:22.532315 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://401a1c1daeea58980316aac659a5f50b88f33db73a751be469657a6bb2849ad4" gracePeriod=600 Oct 02 13:36:22 crc kubenswrapper[4710]: I1002 13:36:22.577009 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-75pwg" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="registry-server" containerID="cri-o://5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f" gracePeriod=2 Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.030578 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.188154 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfvxs\" (UniqueName: \"kubernetes.io/projected/a9cb4718-1ed0-40d9-8378-b3221ce3d529-kube-api-access-kfvxs\") pod \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.188323 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-utilities\") pod \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.188390 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-catalog-content\") pod \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\" (UID: \"a9cb4718-1ed0-40d9-8378-b3221ce3d529\") " Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.189060 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-utilities" (OuterVolumeSpecName: "utilities") pod "a9cb4718-1ed0-40d9-8378-b3221ce3d529" (UID: "a9cb4718-1ed0-40d9-8378-b3221ce3d529"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.206596 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9cb4718-1ed0-40d9-8378-b3221ce3d529-kube-api-access-kfvxs" (OuterVolumeSpecName: "kube-api-access-kfvxs") pod "a9cb4718-1ed0-40d9-8378-b3221ce3d529" (UID: "a9cb4718-1ed0-40d9-8378-b3221ce3d529"). InnerVolumeSpecName "kube-api-access-kfvxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.290766 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.291062 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfvxs\" (UniqueName: \"kubernetes.io/projected/a9cb4718-1ed0-40d9-8378-b3221ce3d529-kube-api-access-kfvxs\") on node \"crc\" DevicePath \"\"" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.591310 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="401a1c1daeea58980316aac659a5f50b88f33db73a751be469657a6bb2849ad4" exitCode=0 Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.591351 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"401a1c1daeea58980316aac659a5f50b88f33db73a751be469657a6bb2849ad4"} Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.591385 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572"} Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.591403 4710 scope.go:117] "RemoveContainer" containerID="ccdec0a71f2c35e71b1aa4b2299a1dce0271d2d207061b347ce40bb46dc3c801" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.595796 4710 generic.go:334] "Generic (PLEG): container finished" podID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerID="5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f" exitCode=0 Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.595831 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerDied","Data":"5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f"} Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.595858 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75pwg" event={"ID":"a9cb4718-1ed0-40d9-8378-b3221ce3d529","Type":"ContainerDied","Data":"caae8e517b7caba82a4f17c5c02cc1041892db938bf4d35d2a791bc38e87ae5c"} Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.595927 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75pwg" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.642980 4710 scope.go:117] "RemoveContainer" containerID="5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.669928 4710 scope.go:117] "RemoveContainer" containerID="01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.695169 4710 scope.go:117] "RemoveContainer" containerID="82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.759488 4710 scope.go:117] "RemoveContainer" containerID="5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f" Oct 02 13:36:23 crc kubenswrapper[4710]: E1002 13:36:23.759888 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f\": container with ID starting with 5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f not found: ID does not exist" containerID="5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.759916 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f"} err="failed to get container status \"5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f\": rpc error: code = NotFound desc = could not find container \"5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f\": container with ID starting with 5673cc2685e70d3abfb3077a370e119d5485a11bdcf16b38b3220087386d557f not found: ID does not exist" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.759935 4710 scope.go:117] "RemoveContainer" containerID="01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05" Oct 02 13:36:23 crc kubenswrapper[4710]: E1002 13:36:23.760284 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05\": container with ID starting with 01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05 not found: ID does not exist" containerID="01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.760339 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05"} err="failed to get container status \"01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05\": rpc error: code = NotFound desc = could not find container \"01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05\": container with ID starting with 01c60a242285b1af97f2cc708803dcbfcbf1419c55e9053d628ceeaca1211f05 not found: ID does not exist" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.760399 4710 scope.go:117] "RemoveContainer" containerID="82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111" Oct 02 13:36:23 crc kubenswrapper[4710]: E1002 13:36:23.760954 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111\": container with ID starting with 82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111 not found: ID does not exist" containerID="82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.760981 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111"} err="failed to get container status \"82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111\": rpc error: code = NotFound desc = could not find container \"82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111\": container with ID starting with 82332d159518669adbe19cc18b36563e03715b791b441ac9c17d81ef7f1fa111 not found: ID does not exist" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.894916 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9cb4718-1ed0-40d9-8378-b3221ce3d529" (UID: "a9cb4718-1ed0-40d9-8378-b3221ce3d529"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.901791 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9cb4718-1ed0-40d9-8378-b3221ce3d529-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.939628 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-75pwg"] Oct 02 13:36:23 crc kubenswrapper[4710]: I1002 13:36:23.953005 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-75pwg"] Oct 02 13:36:24 crc kubenswrapper[4710]: I1002 13:36:24.917173 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" path="/var/lib/kubelet/pods/a9cb4718-1ed0-40d9-8378-b3221ce3d529/volumes" Oct 02 13:38:22 crc kubenswrapper[4710]: I1002 13:38:22.530979 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:38:22 crc kubenswrapper[4710]: I1002 13:38:22.531479 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:38:34 crc kubenswrapper[4710]: I1002 13:38:34.993920 4710 generic.go:334] "Generic (PLEG): container finished" podID="76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" containerID="793f842deb770c4035ba8c3fd5d96b32608969a94124a574915d77b78f39953d" exitCode=0 Oct 02 13:38:34 crc kubenswrapper[4710]: I1002 13:38:34.994016 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" event={"ID":"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8","Type":"ContainerDied","Data":"793f842deb770c4035ba8c3fd5d96b32608969a94124a574915d77b78f39953d"} Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.507287 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613216 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-inventory\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613292 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-0\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613333 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-telemetry-combined-ca-bundle\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613356 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-2\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613404 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-1\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613447 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ssh-key\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.613494 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnwlc\" (UniqueName: \"kubernetes.io/projected/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-kube-api-access-jnwlc\") pod \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\" (UID: \"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8\") " Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.619730 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.623402 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-kube-api-access-jnwlc" (OuterVolumeSpecName: "kube-api-access-jnwlc") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "kube-api-access-jnwlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.650008 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-inventory" (OuterVolumeSpecName: "inventory") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.652785 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.656153 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.668013 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.668437 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" (UID: "76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715251 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715294 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnwlc\" (UniqueName: \"kubernetes.io/projected/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-kube-api-access-jnwlc\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715311 4710 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715324 4710 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715334 4710 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715344 4710 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:36 crc kubenswrapper[4710]: I1002 13:38:36.715355 4710 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 13:38:37 crc kubenswrapper[4710]: I1002 13:38:37.013690 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" event={"ID":"76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8","Type":"ContainerDied","Data":"48a9faa8fca8a4c694569c28293d69933efc744702602cc5462e0a3a21611bb6"} Oct 02 13:38:37 crc kubenswrapper[4710]: I1002 13:38:37.013743 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48a9faa8fca8a4c694569c28293d69933efc744702602cc5462e0a3a21611bb6" Oct 02 13:38:37 crc kubenswrapper[4710]: I1002 13:38:37.013841 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h" Oct 02 13:38:52 crc kubenswrapper[4710]: I1002 13:38:52.530109 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:38:52 crc kubenswrapper[4710]: I1002 13:38:52.530624 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.628963 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b5n6g"] Oct 02 13:38:56 crc kubenswrapper[4710]: E1002 13:38:56.630701 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="extract-utilities" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.630716 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="extract-utilities" Oct 02 13:38:56 crc kubenswrapper[4710]: E1002 13:38:56.630727 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="extract-content" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.630733 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="extract-content" Oct 02 13:38:56 crc kubenswrapper[4710]: E1002 13:38:56.630771 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.630781 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 13:38:56 crc kubenswrapper[4710]: E1002 13:38:56.630792 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="registry-server" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.630797 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="registry-server" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.630997 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.631016 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cb4718-1ed0-40d9-8378-b3221ce3d529" containerName="registry-server" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.632518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.703991 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b5n6g"] Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.723728 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-catalog-content\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.723900 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltr5c\" (UniqueName: \"kubernetes.io/projected/3f05c130-8b39-4721-b5a6-660bcef775ac-kube-api-access-ltr5c\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.723956 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-utilities\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.825238 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-utilities\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.825396 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-catalog-content\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.825455 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltr5c\" (UniqueName: \"kubernetes.io/projected/3f05c130-8b39-4721-b5a6-660bcef775ac-kube-api-access-ltr5c\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.825722 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-utilities\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.825796 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-catalog-content\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.845339 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltr5c\" (UniqueName: \"kubernetes.io/projected/3f05c130-8b39-4721-b5a6-660bcef775ac-kube-api-access-ltr5c\") pod \"community-operators-b5n6g\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:56 crc kubenswrapper[4710]: I1002 13:38:56.967521 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:38:57 crc kubenswrapper[4710]: I1002 13:38:57.530215 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b5n6g"] Oct 02 13:38:58 crc kubenswrapper[4710]: I1002 13:38:58.227115 4710 generic.go:334] "Generic (PLEG): container finished" podID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerID="c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de" exitCode=0 Oct 02 13:38:58 crc kubenswrapper[4710]: I1002 13:38:58.227175 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerDied","Data":"c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de"} Oct 02 13:38:58 crc kubenswrapper[4710]: I1002 13:38:58.227464 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerStarted","Data":"2226891d291f38dc5fb7fea72388fc6a0c181c4e577501c5a1d2f49cce7958bb"} Oct 02 13:38:58 crc kubenswrapper[4710]: I1002 13:38:58.230261 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:38:59 crc kubenswrapper[4710]: I1002 13:38:59.237467 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerStarted","Data":"720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1"} Oct 02 13:39:00 crc kubenswrapper[4710]: I1002 13:39:00.247838 4710 generic.go:334] "Generic (PLEG): container finished" podID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerID="720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1" exitCode=0 Oct 02 13:39:00 crc kubenswrapper[4710]: I1002 13:39:00.247876 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerDied","Data":"720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1"} Oct 02 13:39:01 crc kubenswrapper[4710]: I1002 13:39:01.260160 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerStarted","Data":"06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda"} Oct 02 13:39:01 crc kubenswrapper[4710]: I1002 13:39:01.285310 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b5n6g" podStartSLOduration=2.818637852 podStartE2EDuration="5.285293208s" podCreationTimestamp="2025-10-02 13:38:56 +0000 UTC" firstStartedPulling="2025-10-02 13:38:58.229977499 +0000 UTC m=+2982.336388382" lastFinishedPulling="2025-10-02 13:39:00.696632855 +0000 UTC m=+2984.803043738" observedRunningTime="2025-10-02 13:39:01.277174873 +0000 UTC m=+2985.383585756" watchObservedRunningTime="2025-10-02 13:39:01.285293208 +0000 UTC m=+2985.391704091" Oct 02 13:39:06 crc kubenswrapper[4710]: I1002 13:39:06.968777 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:39:06 crc kubenswrapper[4710]: I1002 13:39:06.969409 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:39:07 crc kubenswrapper[4710]: I1002 13:39:07.019318 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:39:07 crc kubenswrapper[4710]: I1002 13:39:07.374366 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:39:08 crc kubenswrapper[4710]: I1002 13:39:08.213310 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b5n6g"] Oct 02 13:39:09 crc kubenswrapper[4710]: I1002 13:39:09.338008 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b5n6g" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="registry-server" containerID="cri-o://06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda" gracePeriod=2 Oct 02 13:39:09 crc kubenswrapper[4710]: I1002 13:39:09.869480 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.036310 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-catalog-content\") pod \"3f05c130-8b39-4721-b5a6-660bcef775ac\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.036962 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-utilities\") pod \"3f05c130-8b39-4721-b5a6-660bcef775ac\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.037109 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltr5c\" (UniqueName: \"kubernetes.io/projected/3f05c130-8b39-4721-b5a6-660bcef775ac-kube-api-access-ltr5c\") pod \"3f05c130-8b39-4721-b5a6-660bcef775ac\" (UID: \"3f05c130-8b39-4721-b5a6-660bcef775ac\") " Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.038211 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-utilities" (OuterVolumeSpecName: "utilities") pod "3f05c130-8b39-4721-b5a6-660bcef775ac" (UID: "3f05c130-8b39-4721-b5a6-660bcef775ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.038555 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.043460 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f05c130-8b39-4721-b5a6-660bcef775ac-kube-api-access-ltr5c" (OuterVolumeSpecName: "kube-api-access-ltr5c") pod "3f05c130-8b39-4721-b5a6-660bcef775ac" (UID: "3f05c130-8b39-4721-b5a6-660bcef775ac"). InnerVolumeSpecName "kube-api-access-ltr5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.144771 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltr5c\" (UniqueName: \"kubernetes.io/projected/3f05c130-8b39-4721-b5a6-660bcef775ac-kube-api-access-ltr5c\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.347630 4710 generic.go:334] "Generic (PLEG): container finished" podID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerID="06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda" exitCode=0 Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.347676 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerDied","Data":"06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda"} Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.347708 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b5n6g" event={"ID":"3f05c130-8b39-4721-b5a6-660bcef775ac","Type":"ContainerDied","Data":"2226891d291f38dc5fb7fea72388fc6a0c181c4e577501c5a1d2f49cce7958bb"} Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.347726 4710 scope.go:117] "RemoveContainer" containerID="06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.347729 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b5n6g" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.372201 4710 scope.go:117] "RemoveContainer" containerID="720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.391040 4710 scope.go:117] "RemoveContainer" containerID="c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.443677 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f05c130-8b39-4721-b5a6-660bcef775ac" (UID: "3f05c130-8b39-4721-b5a6-660bcef775ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.450356 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f05c130-8b39-4721-b5a6-660bcef775ac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.471513 4710 scope.go:117] "RemoveContainer" containerID="06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda" Oct 02 13:39:10 crc kubenswrapper[4710]: E1002 13:39:10.481683 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda\": container with ID starting with 06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda not found: ID does not exist" containerID="06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.481724 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda"} err="failed to get container status \"06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda\": rpc error: code = NotFound desc = could not find container \"06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda\": container with ID starting with 06932d03d085b35aaf739a8a324386508ebfba4cc8cd7b15f711d1e421fe5bda not found: ID does not exist" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.481758 4710 scope.go:117] "RemoveContainer" containerID="720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1" Oct 02 13:39:10 crc kubenswrapper[4710]: E1002 13:39:10.482147 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1\": container with ID starting with 720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1 not found: ID does not exist" containerID="720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.482185 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1"} err="failed to get container status \"720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1\": rpc error: code = NotFound desc = could not find container \"720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1\": container with ID starting with 720c4cd77f4c44b282a832309c68c13414cd160d9493403ed9ad45158a71cad1 not found: ID does not exist" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.482199 4710 scope.go:117] "RemoveContainer" containerID="c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de" Oct 02 13:39:10 crc kubenswrapper[4710]: E1002 13:39:10.482468 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de\": container with ID starting with c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de not found: ID does not exist" containerID="c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.482495 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de"} err="failed to get container status \"c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de\": rpc error: code = NotFound desc = could not find container \"c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de\": container with ID starting with c09126b40fc9c65ca07db12d6a56c6bd62bc74b9786fe75aa93ffd3025dd04de not found: ID does not exist" Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.683050 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b5n6g"] Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.702599 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b5n6g"] Oct 02 13:39:10 crc kubenswrapper[4710]: I1002 13:39:10.916248 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" path="/var/lib/kubelet/pods/3f05c130-8b39-4721-b5a6-660bcef775ac/volumes" Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.189832 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.190476 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="prometheus" containerID="cri-o://c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88" gracePeriod=600 Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.190891 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="thanos-sidecar" containerID="cri-o://5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271" gracePeriod=600 Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.190932 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="config-reloader" containerID="cri-o://334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0" gracePeriod=600 Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.415269 4710 generic.go:334] "Generic (PLEG): container finished" podID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerID="5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271" exitCode=0 Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.415613 4710 generic.go:334] "Generic (PLEG): container finished" podID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerID="334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0" exitCode=0 Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.415640 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerDied","Data":"5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271"} Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.415689 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerDied","Data":"334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0"} Oct 02 13:39:16 crc kubenswrapper[4710]: I1002 13:39:16.703646 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.137:9090/-/ready\": dial tcp 10.217.0.137:9090: connect: connection refused" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.297109 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.427360 4710 generic.go:334] "Generic (PLEG): container finished" podID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerID="c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88" exitCode=0 Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.427403 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerDied","Data":"c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88"} Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.427427 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e","Type":"ContainerDied","Data":"ecf2c4b319af9b9600dc61fbf2cadcd99efdb00314a72310f645c7e8374462a3"} Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.427447 4710 scope.go:117] "RemoveContainer" containerID="5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.427604 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.454216 4710 scope.go:117] "RemoveContainer" containerID="334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.475178 4710 scope.go:117] "RemoveContainer" containerID="c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493138 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-thanos-prometheus-http-client-file\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493185 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493254 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmkcj\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-kube-api-access-mmkcj\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493286 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config-out\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493345 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493403 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-tls-assets\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493437 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493506 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-secret-combined-ca-bundle\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493534 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-prometheus-metric-storage-rulefiles-0\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493566 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.493662 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\" (UID: \"2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e\") " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.494524 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.494578 4710 scope.go:117] "RemoveContainer" containerID="9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.501785 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.501847 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.502884 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.503156 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.504617 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config" (OuterVolumeSpecName: "config") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.504930 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config-out" (OuterVolumeSpecName: "config-out") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.505270 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-kube-api-access-mmkcj" (OuterVolumeSpecName: "kube-api-access-mmkcj") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "kube-api-access-mmkcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.505388 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.521320 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "pvc-11c95985-a270-4e8f-9943-0f6910c78dc6". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.574648 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config" (OuterVolumeSpecName: "web-config") pod "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" (UID: "2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595396 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmkcj\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-kube-api-access-mmkcj\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595433 4710 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config-out\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595450 4710 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595463 4710 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595476 4710 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595487 4710 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595498 4710 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595511 4710 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595544 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") on node \"crc\" " Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595558 4710 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.595570 4710 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e-web-config\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.625886 4710 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.626033 4710 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-11c95985-a270-4e8f-9943-0f6910c78dc6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6") on node "crc" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.697181 4710 reconciler_common.go:293] "Volume detached for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") on node \"crc\" DevicePath \"\"" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.710883 4710 scope.go:117] "RemoveContainer" containerID="5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.712467 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271\": container with ID starting with 5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271 not found: ID does not exist" containerID="5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.712534 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271"} err="failed to get container status \"5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271\": rpc error: code = NotFound desc = could not find container \"5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271\": container with ID starting with 5c4e37f3378ef1cf833ea76307e9d6898122c92fcb2c21dba02836482b02a271 not found: ID does not exist" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.712573 4710 scope.go:117] "RemoveContainer" containerID="334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.712937 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0\": container with ID starting with 334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0 not found: ID does not exist" containerID="334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.713004 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0"} err="failed to get container status \"334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0\": rpc error: code = NotFound desc = could not find container \"334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0\": container with ID starting with 334a86536b2d8fbd0612a4e61d1cc4ee32ae8670162d33535d47da35055433e0 not found: ID does not exist" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.713031 4710 scope.go:117] "RemoveContainer" containerID="c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.713332 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88\": container with ID starting with c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88 not found: ID does not exist" containerID="c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.713381 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88"} err="failed to get container status \"c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88\": rpc error: code = NotFound desc = could not find container \"c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88\": container with ID starting with c3f6ebdd4e3de61286649533008b0b4b77a3419adba349f4873bac0a9e979c88 not found: ID does not exist" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.713407 4710 scope.go:117] "RemoveContainer" containerID="9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.713760 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6\": container with ID starting with 9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6 not found: ID does not exist" containerID="9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.713799 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6"} err="failed to get container status \"9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6\": rpc error: code = NotFound desc = could not find container \"9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6\": container with ID starting with 9794ec9e734226cfaf1a9400701e774eedb73bb9e7766793d34b19033e6328c6 not found: ID does not exist" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.763611 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.771656 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.787897 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788303 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="registry-server" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788323 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="registry-server" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788530 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="init-config-reloader" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788539 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="init-config-reloader" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788552 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="thanos-sidecar" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788558 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="thanos-sidecar" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788569 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="extract-content" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788574 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="extract-content" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788592 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="prometheus" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788598 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="prometheus" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788614 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="extract-utilities" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788622 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="extract-utilities" Oct 02 13:39:17 crc kubenswrapper[4710]: E1002 13:39:17.788636 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="config-reloader" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788643 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="config-reloader" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788837 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="thanos-sidecar" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788848 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f05c130-8b39-4721-b5a6-660bcef775ac" containerName="registry-server" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788862 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="prometheus" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.788886 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" containerName="config-reloader" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.790849 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.795206 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.795270 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.795458 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.795882 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.797956 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-58df2" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.806001 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.815593 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899528 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899599 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899659 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899680 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899719 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899824 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.899878 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs5g2\" (UniqueName: \"kubernetes.io/projected/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-kube-api-access-qs5g2\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.900086 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.900252 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.900319 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:17 crc kubenswrapper[4710]: I1002 13:39:17.900498 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002180 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002245 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002311 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002346 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002372 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002406 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002435 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002476 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs5g2\" (UniqueName: \"kubernetes.io/projected/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-kube-api-access-qs5g2\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002576 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002687 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.002730 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.005966 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.006737 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.007271 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.007583 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.009593 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.009980 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.011592 4710 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.011620 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/549ec1f26b5b9c00df36a940184c18d591918f74b3d89fea9d49d0eb0309560f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.012196 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.016053 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.028049 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.029546 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs5g2\" (UniqueName: \"kubernetes.io/projected/4c72cadc-76e1-4e5e-aff8-dec25f7c3283-kube-api-access-qs5g2\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.072095 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11c95985-a270-4e8f-9943-0f6910c78dc6\") pod \"prometheus-metric-storage-0\" (UID: \"4c72cadc-76e1-4e5e-aff8-dec25f7c3283\") " pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.114327 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.620360 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 13:39:18 crc kubenswrapper[4710]: I1002 13:39:18.916705 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e" path="/var/lib/kubelet/pods/2f08e1f9-9d39-4b16-8c2e-eaf3c1bbca5e/volumes" Oct 02 13:39:19 crc kubenswrapper[4710]: I1002 13:39:19.448533 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c72cadc-76e1-4e5e-aff8-dec25f7c3283","Type":"ContainerStarted","Data":"1ba44dc14062d9a7249670d322eadfb32ff4820d7b82dea5a82f6acfa112e930"} Oct 02 13:39:22 crc kubenswrapper[4710]: I1002 13:39:22.479976 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c72cadc-76e1-4e5e-aff8-dec25f7c3283","Type":"ContainerStarted","Data":"809f22431e20482ee7b7240d9b44c43fe61a78586317108904b80af14b52cc4c"} Oct 02 13:39:22 crc kubenswrapper[4710]: I1002 13:39:22.530531 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:39:22 crc kubenswrapper[4710]: I1002 13:39:22.530592 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:39:22 crc kubenswrapper[4710]: I1002 13:39:22.530635 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:39:22 crc kubenswrapper[4710]: I1002 13:39:22.531353 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:39:22 crc kubenswrapper[4710]: I1002 13:39:22.531407 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" gracePeriod=600 Oct 02 13:39:22 crc kubenswrapper[4710]: E1002 13:39:22.655689 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:39:23 crc kubenswrapper[4710]: I1002 13:39:23.495274 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" exitCode=0 Oct 02 13:39:23 crc kubenswrapper[4710]: I1002 13:39:23.495370 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572"} Oct 02 13:39:23 crc kubenswrapper[4710]: I1002 13:39:23.495723 4710 scope.go:117] "RemoveContainer" containerID="401a1c1daeea58980316aac659a5f50b88f33db73a751be469657a6bb2849ad4" Oct 02 13:39:23 crc kubenswrapper[4710]: I1002 13:39:23.496649 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:39:23 crc kubenswrapper[4710]: E1002 13:39:23.497044 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:39:29 crc kubenswrapper[4710]: I1002 13:39:29.563570 4710 generic.go:334] "Generic (PLEG): container finished" podID="4c72cadc-76e1-4e5e-aff8-dec25f7c3283" containerID="809f22431e20482ee7b7240d9b44c43fe61a78586317108904b80af14b52cc4c" exitCode=0 Oct 02 13:39:29 crc kubenswrapper[4710]: I1002 13:39:29.563753 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c72cadc-76e1-4e5e-aff8-dec25f7c3283","Type":"ContainerDied","Data":"809f22431e20482ee7b7240d9b44c43fe61a78586317108904b80af14b52cc4c"} Oct 02 13:39:30 crc kubenswrapper[4710]: I1002 13:39:30.575175 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c72cadc-76e1-4e5e-aff8-dec25f7c3283","Type":"ContainerStarted","Data":"5c8ab7cb1e806c33a45fdf58720efbddfe1d368514bc3db6955019a0ac58544f"} Oct 02 13:39:33 crc kubenswrapper[4710]: I1002 13:39:33.604240 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c72cadc-76e1-4e5e-aff8-dec25f7c3283","Type":"ContainerStarted","Data":"440980ac681ccc5bc41765520a5724695e3d3ee2f13c5ae65cab03c982f7e120"} Oct 02 13:39:33 crc kubenswrapper[4710]: I1002 13:39:33.604811 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4c72cadc-76e1-4e5e-aff8-dec25f7c3283","Type":"ContainerStarted","Data":"bfa4f395daf842c1c6ee6441f5a44e4cd2e9a510f7066ec6e0f7e4a1e2dcf717"} Oct 02 13:39:35 crc kubenswrapper[4710]: I1002 13:39:35.905551 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:39:35 crc kubenswrapper[4710]: E1002 13:39:35.906448 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:39:38 crc kubenswrapper[4710]: I1002 13:39:38.114518 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:46 crc kubenswrapper[4710]: I1002 13:39:46.924983 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:39:46 crc kubenswrapper[4710]: E1002 13:39:46.926502 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:39:48 crc kubenswrapper[4710]: I1002 13:39:48.115179 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:48 crc kubenswrapper[4710]: I1002 13:39:48.124416 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:48 crc kubenswrapper[4710]: I1002 13:39:48.157257 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=31.157228938 podStartE2EDuration="31.157228938s" podCreationTimestamp="2025-10-02 13:39:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:39:33.632578526 +0000 UTC m=+3017.738989409" watchObservedRunningTime="2025-10-02 13:39:48.157228938 +0000 UTC m=+3032.263639851" Oct 02 13:39:48 crc kubenswrapper[4710]: I1002 13:39:48.766858 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 13:39:57 crc kubenswrapper[4710]: I1002 13:39:57.904409 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:39:57 crc kubenswrapper[4710]: E1002 13:39:57.905175 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.530723 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.533298 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.535858 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.536079 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pbnjd" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.536332 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.536636 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.548707 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.641791 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.641855 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.641894 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.641934 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28gwc\" (UniqueName: \"kubernetes.io/projected/22910690-8789-41f2-bd0a-b56b2c6683a6-kube-api-access-28gwc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.642214 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-config-data\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.642332 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.642367 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.642425 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.642608 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745432 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745490 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745510 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28gwc\" (UniqueName: \"kubernetes.io/projected/22910690-8789-41f2-bd0a-b56b2c6683a6-kube-api-access-28gwc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745577 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-config-data\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745616 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745636 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745665 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745720 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.745758 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.746114 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.746378 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.746461 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.751554 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.752217 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.752849 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.754623 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-config-data\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.760352 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.764730 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28gwc\" (UniqueName: \"kubernetes.io/projected/22910690-8789-41f2-bd0a-b56b2c6683a6-kube-api-access-28gwc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.776022 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " pod="openstack/tempest-tests-tempest" Oct 02 13:40:08 crc kubenswrapper[4710]: I1002 13:40:08.871229 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 13:40:09 crc kubenswrapper[4710]: I1002 13:40:09.865613 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 13:40:10 crc kubenswrapper[4710]: I1002 13:40:10.005337 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"22910690-8789-41f2-bd0a-b56b2c6683a6","Type":"ContainerStarted","Data":"2949348ca1c21d04b0748975a3208024671a98e42dcfb287ae00a778a3847734"} Oct 02 13:40:10 crc kubenswrapper[4710]: I1002 13:40:10.911058 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:40:10 crc kubenswrapper[4710]: E1002 13:40:10.915404 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:40:20 crc kubenswrapper[4710]: I1002 13:40:20.801266 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 13:40:22 crc kubenswrapper[4710]: I1002 13:40:22.121231 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"22910690-8789-41f2-bd0a-b56b2c6683a6","Type":"ContainerStarted","Data":"66439e1c1481f3a465c44d5c66708295824bc9dea2c8ec2d4e218a74474708e8"} Oct 02 13:40:24 crc kubenswrapper[4710]: I1002 13:40:24.904831 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:40:24 crc kubenswrapper[4710]: E1002 13:40:24.905552 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:40:39 crc kubenswrapper[4710]: I1002 13:40:39.905599 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:40:39 crc kubenswrapper[4710]: E1002 13:40:39.906345 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:40:47 crc kubenswrapper[4710]: I1002 13:40:47.607844 4710 scope.go:117] "RemoveContainer" containerID="02c64043dd7791e2877c1a2340d01e2c3d4639793891a06252b50b22bad05588" Oct 02 13:40:47 crc kubenswrapper[4710]: I1002 13:40:47.638638 4710 scope.go:117] "RemoveContainer" containerID="2c70cc29ffe895201925454d271645ee9082e769c35ae836a2432ef5b240a8cb" Oct 02 13:40:47 crc kubenswrapper[4710]: I1002 13:40:47.687728 4710 scope.go:117] "RemoveContainer" containerID="89ccb492130b67f244cf132665d90be94c61da388925dc7d88417a2126debc57" Oct 02 13:40:51 crc kubenswrapper[4710]: I1002 13:40:51.905218 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:40:51 crc kubenswrapper[4710]: E1002 13:40:51.906066 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:41:02 crc kubenswrapper[4710]: I1002 13:41:02.905784 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:41:02 crc kubenswrapper[4710]: E1002 13:41:02.906919 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:41:17 crc kubenswrapper[4710]: I1002 13:41:17.905224 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:41:17 crc kubenswrapper[4710]: E1002 13:41:17.905966 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:41:32 crc kubenswrapper[4710]: I1002 13:41:32.905334 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:41:32 crc kubenswrapper[4710]: E1002 13:41:32.906644 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:41:46 crc kubenswrapper[4710]: I1002 13:41:46.913740 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:41:46 crc kubenswrapper[4710]: E1002 13:41:46.914531 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:41:58 crc kubenswrapper[4710]: I1002 13:41:58.904868 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:41:58 crc kubenswrapper[4710]: E1002 13:41:58.905777 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:42:11 crc kubenswrapper[4710]: I1002 13:42:11.904707 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:42:11 crc kubenswrapper[4710]: E1002 13:42:11.905364 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:42:26 crc kubenswrapper[4710]: I1002 13:42:26.916290 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:42:26 crc kubenswrapper[4710]: E1002 13:42:26.917045 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:42:40 crc kubenswrapper[4710]: I1002 13:42:40.904412 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:42:40 crc kubenswrapper[4710]: E1002 13:42:40.905813 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:42:53 crc kubenswrapper[4710]: I1002 13:42:53.905047 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:42:53 crc kubenswrapper[4710]: E1002 13:42:53.905873 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:43:05 crc kubenswrapper[4710]: I1002 13:43:05.904280 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:43:05 crc kubenswrapper[4710]: E1002 13:43:05.905097 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:43:19 crc kubenswrapper[4710]: I1002 13:43:19.905673 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:43:19 crc kubenswrapper[4710]: E1002 13:43:19.906427 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:43:31 crc kubenswrapper[4710]: I1002 13:43:31.904840 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:43:31 crc kubenswrapper[4710]: E1002 13:43:31.905571 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:43:42 crc kubenswrapper[4710]: I1002 13:43:42.904837 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:43:42 crc kubenswrapper[4710]: E1002 13:43:42.905862 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:43:53 crc kubenswrapper[4710]: I1002 13:43:53.904628 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:43:53 crc kubenswrapper[4710]: E1002 13:43:53.906624 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:44:08 crc kubenswrapper[4710]: I1002 13:44:08.904658 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:44:08 crc kubenswrapper[4710]: E1002 13:44:08.906031 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:44:22 crc kubenswrapper[4710]: I1002 13:44:22.905042 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:44:23 crc kubenswrapper[4710]: I1002 13:44:23.784422 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"0ac006c31e441a8e868caaced27226ba6c8e96c7722cf7086f6c76445baa2a8f"} Oct 02 13:44:23 crc kubenswrapper[4710]: I1002 13:44:23.820469 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=245.891007911 podStartE2EDuration="4m16.820454862s" podCreationTimestamp="2025-10-02 13:40:07 +0000 UTC" firstStartedPulling="2025-10-02 13:40:09.868695446 +0000 UTC m=+3053.975106359" lastFinishedPulling="2025-10-02 13:40:20.798142427 +0000 UTC m=+3064.904553310" observedRunningTime="2025-10-02 13:40:22.147245699 +0000 UTC m=+3066.253656582" watchObservedRunningTime="2025-10-02 13:44:23.820454862 +0000 UTC m=+3307.926865745" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.441950 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v2pts"] Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.451111 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.458055 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2pts"] Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.596952 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-utilities\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.597052 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-catalog-content\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.597295 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlvdj\" (UniqueName: \"kubernetes.io/projected/358cb7bb-ef97-4107-bff1-14f337f231a4-kube-api-access-mlvdj\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.699829 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-catalog-content\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.700292 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlvdj\" (UniqueName: \"kubernetes.io/projected/358cb7bb-ef97-4107-bff1-14f337f231a4-kube-api-access-mlvdj\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.700383 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-catalog-content\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.700668 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-utilities\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.700949 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-utilities\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.725923 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlvdj\" (UniqueName: \"kubernetes.io/projected/358cb7bb-ef97-4107-bff1-14f337f231a4-kube-api-access-mlvdj\") pod \"redhat-marketplace-v2pts\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:33 crc kubenswrapper[4710]: I1002 13:44:33.782226 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:34 crc kubenswrapper[4710]: I1002 13:44:34.247531 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2pts"] Oct 02 13:44:34 crc kubenswrapper[4710]: I1002 13:44:34.898024 4710 generic.go:334] "Generic (PLEG): container finished" podID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerID="f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e" exitCode=0 Oct 02 13:44:34 crc kubenswrapper[4710]: I1002 13:44:34.898077 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2pts" event={"ID":"358cb7bb-ef97-4107-bff1-14f337f231a4","Type":"ContainerDied","Data":"f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e"} Oct 02 13:44:34 crc kubenswrapper[4710]: I1002 13:44:34.898375 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2pts" event={"ID":"358cb7bb-ef97-4107-bff1-14f337f231a4","Type":"ContainerStarted","Data":"8dd606dc6d16dc5627bfd47d56e4636e25c8a18d2ee7deb9c1bdbe916d858e78"} Oct 02 13:44:34 crc kubenswrapper[4710]: I1002 13:44:34.900579 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:44:35 crc kubenswrapper[4710]: I1002 13:44:35.920712 4710 generic.go:334] "Generic (PLEG): container finished" podID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerID="e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06" exitCode=0 Oct 02 13:44:35 crc kubenswrapper[4710]: I1002 13:44:35.920794 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2pts" event={"ID":"358cb7bb-ef97-4107-bff1-14f337f231a4","Type":"ContainerDied","Data":"e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06"} Oct 02 13:44:37 crc kubenswrapper[4710]: I1002 13:44:37.947018 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2pts" event={"ID":"358cb7bb-ef97-4107-bff1-14f337f231a4","Type":"ContainerStarted","Data":"efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668"} Oct 02 13:44:37 crc kubenswrapper[4710]: I1002 13:44:37.962473 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v2pts" podStartSLOduration=2.957954402 podStartE2EDuration="4.962450629s" podCreationTimestamp="2025-10-02 13:44:33 +0000 UTC" firstStartedPulling="2025-10-02 13:44:34.900275678 +0000 UTC m=+3319.006686571" lastFinishedPulling="2025-10-02 13:44:36.904771915 +0000 UTC m=+3321.011182798" observedRunningTime="2025-10-02 13:44:37.961620208 +0000 UTC m=+3322.068031101" watchObservedRunningTime="2025-10-02 13:44:37.962450629 +0000 UTC m=+3322.068861512" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.621030 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kp6bw"] Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.623420 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.651044 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kp6bw"] Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.724945 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-utilities\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.725013 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-catalog-content\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.725294 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrvc\" (UniqueName: \"kubernetes.io/projected/8fec9da4-3246-4c3e-af10-849d2b3a2596-kube-api-access-jzrvc\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.826764 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrvc\" (UniqueName: \"kubernetes.io/projected/8fec9da4-3246-4c3e-af10-849d2b3a2596-kube-api-access-jzrvc\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.826934 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-utilities\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.826988 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-catalog-content\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.827947 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-catalog-content\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.828044 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-utilities\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.852986 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrvc\" (UniqueName: \"kubernetes.io/projected/8fec9da4-3246-4c3e-af10-849d2b3a2596-kube-api-access-jzrvc\") pod \"redhat-operators-kp6bw\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:39 crc kubenswrapper[4710]: I1002 13:44:39.953726 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:40 crc kubenswrapper[4710]: I1002 13:44:40.428682 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kp6bw"] Oct 02 13:44:40 crc kubenswrapper[4710]: W1002 13:44:40.438180 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fec9da4_3246_4c3e_af10_849d2b3a2596.slice/crio-23bb86e32e8a7575a7a3bfc40715234c94b0002d1d60f581c921db8e4f0b6d98 WatchSource:0}: Error finding container 23bb86e32e8a7575a7a3bfc40715234c94b0002d1d60f581c921db8e4f0b6d98: Status 404 returned error can't find the container with id 23bb86e32e8a7575a7a3bfc40715234c94b0002d1d60f581c921db8e4f0b6d98 Oct 02 13:44:40 crc kubenswrapper[4710]: I1002 13:44:40.981437 4710 generic.go:334] "Generic (PLEG): container finished" podID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerID="4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44" exitCode=0 Oct 02 13:44:40 crc kubenswrapper[4710]: I1002 13:44:40.981489 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerDied","Data":"4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44"} Oct 02 13:44:40 crc kubenswrapper[4710]: I1002 13:44:40.981893 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerStarted","Data":"23bb86e32e8a7575a7a3bfc40715234c94b0002d1d60f581c921db8e4f0b6d98"} Oct 02 13:44:41 crc kubenswrapper[4710]: I1002 13:44:41.996026 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerStarted","Data":"dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f"} Oct 02 13:44:43 crc kubenswrapper[4710]: I1002 13:44:43.782604 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:43 crc kubenswrapper[4710]: I1002 13:44:43.782674 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:43 crc kubenswrapper[4710]: I1002 13:44:43.878777 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:44 crc kubenswrapper[4710]: I1002 13:44:44.087376 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:45 crc kubenswrapper[4710]: I1002 13:44:45.012297 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2pts"] Oct 02 13:44:45 crc kubenswrapper[4710]: I1002 13:44:45.026504 4710 generic.go:334] "Generic (PLEG): container finished" podID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerID="dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f" exitCode=0 Oct 02 13:44:45 crc kubenswrapper[4710]: I1002 13:44:45.026540 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerDied","Data":"dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f"} Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.044652 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v2pts" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="registry-server" containerID="cri-o://efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668" gracePeriod=2 Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.044923 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerStarted","Data":"152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646"} Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.508905 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.524305 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kp6bw" podStartSLOduration=2.977177331 podStartE2EDuration="7.524283204s" podCreationTimestamp="2025-10-02 13:44:39 +0000 UTC" firstStartedPulling="2025-10-02 13:44:40.98380668 +0000 UTC m=+3325.090217563" lastFinishedPulling="2025-10-02 13:44:45.530912543 +0000 UTC m=+3329.637323436" observedRunningTime="2025-10-02 13:44:46.073668347 +0000 UTC m=+3330.180079250" watchObservedRunningTime="2025-10-02 13:44:46.524283204 +0000 UTC m=+3330.630694087" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.710087 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlvdj\" (UniqueName: \"kubernetes.io/projected/358cb7bb-ef97-4107-bff1-14f337f231a4-kube-api-access-mlvdj\") pod \"358cb7bb-ef97-4107-bff1-14f337f231a4\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.710355 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-catalog-content\") pod \"358cb7bb-ef97-4107-bff1-14f337f231a4\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.710532 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-utilities\") pod \"358cb7bb-ef97-4107-bff1-14f337f231a4\" (UID: \"358cb7bb-ef97-4107-bff1-14f337f231a4\") " Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.711280 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-utilities" (OuterVolumeSpecName: "utilities") pod "358cb7bb-ef97-4107-bff1-14f337f231a4" (UID: "358cb7bb-ef97-4107-bff1-14f337f231a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.722423 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "358cb7bb-ef97-4107-bff1-14f337f231a4" (UID: "358cb7bb-ef97-4107-bff1-14f337f231a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.722890 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/358cb7bb-ef97-4107-bff1-14f337f231a4-kube-api-access-mlvdj" (OuterVolumeSpecName: "kube-api-access-mlvdj") pod "358cb7bb-ef97-4107-bff1-14f337f231a4" (UID: "358cb7bb-ef97-4107-bff1-14f337f231a4"). InnerVolumeSpecName "kube-api-access-mlvdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.812696 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.812770 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlvdj\" (UniqueName: \"kubernetes.io/projected/358cb7bb-ef97-4107-bff1-14f337f231a4-kube-api-access-mlvdj\") on node \"crc\" DevicePath \"\"" Oct 02 13:44:46 crc kubenswrapper[4710]: I1002 13:44:46.812789 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358cb7bb-ef97-4107-bff1-14f337f231a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.058674 4710 generic.go:334] "Generic (PLEG): container finished" podID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerID="efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668" exitCode=0 Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.058752 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2pts" event={"ID":"358cb7bb-ef97-4107-bff1-14f337f231a4","Type":"ContainerDied","Data":"efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668"} Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.058829 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2pts" event={"ID":"358cb7bb-ef97-4107-bff1-14f337f231a4","Type":"ContainerDied","Data":"8dd606dc6d16dc5627bfd47d56e4636e25c8a18d2ee7deb9c1bdbe916d858e78"} Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.058853 4710 scope.go:117] "RemoveContainer" containerID="efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.059258 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2pts" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.087294 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2pts"] Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.092620 4710 scope.go:117] "RemoveContainer" containerID="e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.096692 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2pts"] Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.113340 4710 scope.go:117] "RemoveContainer" containerID="f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.162923 4710 scope.go:117] "RemoveContainer" containerID="efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668" Oct 02 13:44:47 crc kubenswrapper[4710]: E1002 13:44:47.163369 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668\": container with ID starting with efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668 not found: ID does not exist" containerID="efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.163469 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668"} err="failed to get container status \"efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668\": rpc error: code = NotFound desc = could not find container \"efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668\": container with ID starting with efd6dd604d08f7b318739f3dd6ea420d311f3504803a10dc470e387eab71c668 not found: ID does not exist" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.163548 4710 scope.go:117] "RemoveContainer" containerID="e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06" Oct 02 13:44:47 crc kubenswrapper[4710]: E1002 13:44:47.163846 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06\": container with ID starting with e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06 not found: ID does not exist" containerID="e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.163914 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06"} err="failed to get container status \"e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06\": rpc error: code = NotFound desc = could not find container \"e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06\": container with ID starting with e33669d5ee30c748a77b8544b22707a5ea237f7a997e561965f70527b224aa06 not found: ID does not exist" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.163988 4710 scope.go:117] "RemoveContainer" containerID="f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e" Oct 02 13:44:47 crc kubenswrapper[4710]: E1002 13:44:47.164221 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e\": container with ID starting with f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e not found: ID does not exist" containerID="f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e" Oct 02 13:44:47 crc kubenswrapper[4710]: I1002 13:44:47.164294 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e"} err="failed to get container status \"f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e\": rpc error: code = NotFound desc = could not find container \"f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e\": container with ID starting with f04267641901d644dea5e9c91e6dec3a8c71162300f6f0fdb2fa51afde8bf83e not found: ID does not exist" Oct 02 13:44:48 crc kubenswrapper[4710]: I1002 13:44:48.915663 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" path="/var/lib/kubelet/pods/358cb7bb-ef97-4107-bff1-14f337f231a4/volumes" Oct 02 13:44:49 crc kubenswrapper[4710]: I1002 13:44:49.954583 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:49 crc kubenswrapper[4710]: I1002 13:44:49.954895 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:44:51 crc kubenswrapper[4710]: I1002 13:44:51.016811 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kp6bw" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="registry-server" probeResult="failure" output=< Oct 02 13:44:51 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 13:44:51 crc kubenswrapper[4710]: > Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.047035 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.130689 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.179845 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77"] Oct 02 13:45:00 crc kubenswrapper[4710]: E1002 13:45:00.180319 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="registry-server" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.180335 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="registry-server" Oct 02 13:45:00 crc kubenswrapper[4710]: E1002 13:45:00.180365 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="extract-content" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.180373 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="extract-content" Oct 02 13:45:00 crc kubenswrapper[4710]: E1002 13:45:00.180391 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="extract-utilities" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.180399 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="extract-utilities" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.180676 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="358cb7bb-ef97-4107-bff1-14f337f231a4" containerName="registry-server" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.181518 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.191440 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.192063 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.213297 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ad83ee3-894e-4059-89c7-b1605e4a4b65-config-volume\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.213538 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l47rw\" (UniqueName: \"kubernetes.io/projected/3ad83ee3-894e-4059-89c7-b1605e4a4b65-kube-api-access-l47rw\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.213574 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ad83ee3-894e-4059-89c7-b1605e4a4b65-secret-volume\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.215823 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77"] Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.300406 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kp6bw"] Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.315416 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ad83ee3-894e-4059-89c7-b1605e4a4b65-config-volume\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.315609 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l47rw\" (UniqueName: \"kubernetes.io/projected/3ad83ee3-894e-4059-89c7-b1605e4a4b65-kube-api-access-l47rw\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.315648 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ad83ee3-894e-4059-89c7-b1605e4a4b65-secret-volume\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.316588 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ad83ee3-894e-4059-89c7-b1605e4a4b65-config-volume\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.323637 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ad83ee3-894e-4059-89c7-b1605e4a4b65-secret-volume\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.335184 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l47rw\" (UniqueName: \"kubernetes.io/projected/3ad83ee3-894e-4059-89c7-b1605e4a4b65-kube-api-access-l47rw\") pod \"collect-profiles-29323545-b8k77\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.510832 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:00 crc kubenswrapper[4710]: I1002 13:45:00.997388 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77"] Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.214700 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" event={"ID":"3ad83ee3-894e-4059-89c7-b1605e4a4b65","Type":"ContainerStarted","Data":"7baca12dbfbf17eab225dab9032783139c35b155a94dff89ab7757f3840fabf2"} Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.214767 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" event={"ID":"3ad83ee3-894e-4059-89c7-b1605e4a4b65","Type":"ContainerStarted","Data":"eb99103a7eed3e4fa8a80980916f5a04b38f7719bc4a6fb804f05a0283742622"} Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.214858 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kp6bw" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="registry-server" containerID="cri-o://152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646" gracePeriod=2 Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.237123 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" podStartSLOduration=1.237102536 podStartE2EDuration="1.237102536s" podCreationTimestamp="2025-10-02 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 13:45:01.232509271 +0000 UTC m=+3345.338920164" watchObservedRunningTime="2025-10-02 13:45:01.237102536 +0000 UTC m=+3345.343513419" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.687607 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.843222 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzrvc\" (UniqueName: \"kubernetes.io/projected/8fec9da4-3246-4c3e-af10-849d2b3a2596-kube-api-access-jzrvc\") pod \"8fec9da4-3246-4c3e-af10-849d2b3a2596\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.843589 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-utilities\") pod \"8fec9da4-3246-4c3e-af10-849d2b3a2596\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.843639 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-catalog-content\") pod \"8fec9da4-3246-4c3e-af10-849d2b3a2596\" (UID: \"8fec9da4-3246-4c3e-af10-849d2b3a2596\") " Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.844490 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-utilities" (OuterVolumeSpecName: "utilities") pod "8fec9da4-3246-4c3e-af10-849d2b3a2596" (UID: "8fec9da4-3246-4c3e-af10-849d2b3a2596"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.856526 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fec9da4-3246-4c3e-af10-849d2b3a2596-kube-api-access-jzrvc" (OuterVolumeSpecName: "kube-api-access-jzrvc") pod "8fec9da4-3246-4c3e-af10-849d2b3a2596" (UID: "8fec9da4-3246-4c3e-af10-849d2b3a2596"). InnerVolumeSpecName "kube-api-access-jzrvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.933130 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fec9da4-3246-4c3e-af10-849d2b3a2596" (UID: "8fec9da4-3246-4c3e-af10-849d2b3a2596"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.946474 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.946527 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fec9da4-3246-4c3e-af10-849d2b3a2596-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:45:01 crc kubenswrapper[4710]: I1002 13:45:01.946541 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzrvc\" (UniqueName: \"kubernetes.io/projected/8fec9da4-3246-4c3e-af10-849d2b3a2596-kube-api-access-jzrvc\") on node \"crc\" DevicePath \"\"" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.232369 4710 generic.go:334] "Generic (PLEG): container finished" podID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerID="152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646" exitCode=0 Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.232465 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerDied","Data":"152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646"} Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.232500 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kp6bw" event={"ID":"8fec9da4-3246-4c3e-af10-849d2b3a2596","Type":"ContainerDied","Data":"23bb86e32e8a7575a7a3bfc40715234c94b0002d1d60f581c921db8e4f0b6d98"} Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.232521 4710 scope.go:117] "RemoveContainer" containerID="152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.233945 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kp6bw" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.241218 4710 generic.go:334] "Generic (PLEG): container finished" podID="3ad83ee3-894e-4059-89c7-b1605e4a4b65" containerID="7baca12dbfbf17eab225dab9032783139c35b155a94dff89ab7757f3840fabf2" exitCode=0 Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.241269 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" event={"ID":"3ad83ee3-894e-4059-89c7-b1605e4a4b65","Type":"ContainerDied","Data":"7baca12dbfbf17eab225dab9032783139c35b155a94dff89ab7757f3840fabf2"} Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.277602 4710 scope.go:117] "RemoveContainer" containerID="dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.288651 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kp6bw"] Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.301082 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kp6bw"] Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.301943 4710 scope.go:117] "RemoveContainer" containerID="4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.347871 4710 scope.go:117] "RemoveContainer" containerID="152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646" Oct 02 13:45:02 crc kubenswrapper[4710]: E1002 13:45:02.348365 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646\": container with ID starting with 152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646 not found: ID does not exist" containerID="152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.348419 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646"} err="failed to get container status \"152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646\": rpc error: code = NotFound desc = could not find container \"152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646\": container with ID starting with 152fc40dfdd33f2138dac1a00a9196046c8052a841c829d1c9c743fab86b5646 not found: ID does not exist" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.348457 4710 scope.go:117] "RemoveContainer" containerID="dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f" Oct 02 13:45:02 crc kubenswrapper[4710]: E1002 13:45:02.348844 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f\": container with ID starting with dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f not found: ID does not exist" containerID="dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.348877 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f"} err="failed to get container status \"dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f\": rpc error: code = NotFound desc = could not find container \"dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f\": container with ID starting with dfb177d42d22bc8aab95385fc3ad28bb8833965fe8a2c6f41aab5ff755a1868f not found: ID does not exist" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.348926 4710 scope.go:117] "RemoveContainer" containerID="4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44" Oct 02 13:45:02 crc kubenswrapper[4710]: E1002 13:45:02.349290 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44\": container with ID starting with 4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44 not found: ID does not exist" containerID="4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.349345 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44"} err="failed to get container status \"4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44\": rpc error: code = NotFound desc = could not find container \"4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44\": container with ID starting with 4556ef990a1905067296be8ea1294bcb9441ca18de46f1b60085e659f00f2d44 not found: ID does not exist" Oct 02 13:45:02 crc kubenswrapper[4710]: I1002 13:45:02.926135 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" path="/var/lib/kubelet/pods/8fec9da4-3246-4c3e-af10-849d2b3a2596/volumes" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.630186 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.791837 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l47rw\" (UniqueName: \"kubernetes.io/projected/3ad83ee3-894e-4059-89c7-b1605e4a4b65-kube-api-access-l47rw\") pod \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.792387 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ad83ee3-894e-4059-89c7-b1605e4a4b65-config-volume\") pod \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.792510 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ad83ee3-894e-4059-89c7-b1605e4a4b65-secret-volume\") pod \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\" (UID: \"3ad83ee3-894e-4059-89c7-b1605e4a4b65\") " Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.793011 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ad83ee3-894e-4059-89c7-b1605e4a4b65-config-volume" (OuterVolumeSpecName: "config-volume") pod "3ad83ee3-894e-4059-89c7-b1605e4a4b65" (UID: "3ad83ee3-894e-4059-89c7-b1605e4a4b65"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.793303 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ad83ee3-894e-4059-89c7-b1605e4a4b65-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.801166 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ad83ee3-894e-4059-89c7-b1605e4a4b65-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3ad83ee3-894e-4059-89c7-b1605e4a4b65" (UID: "3ad83ee3-894e-4059-89c7-b1605e4a4b65"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.801218 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad83ee3-894e-4059-89c7-b1605e4a4b65-kube-api-access-l47rw" (OuterVolumeSpecName: "kube-api-access-l47rw") pod "3ad83ee3-894e-4059-89c7-b1605e4a4b65" (UID: "3ad83ee3-894e-4059-89c7-b1605e4a4b65"). InnerVolumeSpecName "kube-api-access-l47rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.895621 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l47rw\" (UniqueName: \"kubernetes.io/projected/3ad83ee3-894e-4059-89c7-b1605e4a4b65-kube-api-access-l47rw\") on node \"crc\" DevicePath \"\"" Oct 02 13:45:03 crc kubenswrapper[4710]: I1002 13:45:03.895676 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ad83ee3-894e-4059-89c7-b1605e4a4b65-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 13:45:04 crc kubenswrapper[4710]: I1002 13:45:04.264280 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" event={"ID":"3ad83ee3-894e-4059-89c7-b1605e4a4b65","Type":"ContainerDied","Data":"eb99103a7eed3e4fa8a80980916f5a04b38f7719bc4a6fb804f05a0283742622"} Oct 02 13:45:04 crc kubenswrapper[4710]: I1002 13:45:04.264342 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb99103a7eed3e4fa8a80980916f5a04b38f7719bc4a6fb804f05a0283742622" Oct 02 13:45:04 crc kubenswrapper[4710]: I1002 13:45:04.264382 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77" Oct 02 13:45:04 crc kubenswrapper[4710]: I1002 13:45:04.321913 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk"] Oct 02 13:45:04 crc kubenswrapper[4710]: I1002 13:45:04.330271 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323500-tzbrk"] Oct 02 13:45:04 crc kubenswrapper[4710]: I1002 13:45:04.920320 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0568be7-153e-4be2-adb6-605c17922339" path="/var/lib/kubelet/pods/b0568be7-153e-4be2-adb6-605c17922339/volumes" Oct 02 13:45:47 crc kubenswrapper[4710]: I1002 13:45:47.895012 4710 scope.go:117] "RemoveContainer" containerID="5743fc5ddd6c575120b5ee40cccf3796d479180679a6cb7edc08ba28568a33fc" Oct 02 13:46:52 crc kubenswrapper[4710]: I1002 13:46:52.530953 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:46:52 crc kubenswrapper[4710]: I1002 13:46:52.531490 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:47:22 crc kubenswrapper[4710]: I1002 13:47:22.529983 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:47:22 crc kubenswrapper[4710]: I1002 13:47:22.530509 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.475149 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8mtxd"] Oct 02 13:47:38 crc kubenswrapper[4710]: E1002 13:47:38.477571 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="extract-utilities" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.477763 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="extract-utilities" Oct 02 13:47:38 crc kubenswrapper[4710]: E1002 13:47:38.477874 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="extract-content" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.477921 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="extract-content" Oct 02 13:47:38 crc kubenswrapper[4710]: E1002 13:47:38.478022 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="registry-server" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.478045 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="registry-server" Oct 02 13:47:38 crc kubenswrapper[4710]: E1002 13:47:38.478102 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad83ee3-894e-4059-89c7-b1605e4a4b65" containerName="collect-profiles" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.478120 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad83ee3-894e-4059-89c7-b1605e4a4b65" containerName="collect-profiles" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.479346 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ad83ee3-894e-4059-89c7-b1605e4a4b65" containerName="collect-profiles" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.479410 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fec9da4-3246-4c3e-af10-849d2b3a2596" containerName="registry-server" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.481291 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.496574 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8mtxd"] Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.586396 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-utilities\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.586568 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-catalog-content\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.586611 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whh47\" (UniqueName: \"kubernetes.io/projected/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-kube-api-access-whh47\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.688310 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whh47\" (UniqueName: \"kubernetes.io/projected/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-kube-api-access-whh47\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.688480 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-utilities\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.688698 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-catalog-content\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.689247 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-utilities\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.689382 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-catalog-content\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.719375 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whh47\" (UniqueName: \"kubernetes.io/projected/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-kube-api-access-whh47\") pod \"certified-operators-8mtxd\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:38 crc kubenswrapper[4710]: I1002 13:47:38.811433 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:39 crc kubenswrapper[4710]: I1002 13:47:39.392404 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8mtxd"] Oct 02 13:47:39 crc kubenswrapper[4710]: I1002 13:47:39.890698 4710 generic.go:334] "Generic (PLEG): container finished" podID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerID="61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7" exitCode=0 Oct 02 13:47:39 crc kubenswrapper[4710]: I1002 13:47:39.890775 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerDied","Data":"61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7"} Oct 02 13:47:39 crc kubenswrapper[4710]: I1002 13:47:39.890828 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerStarted","Data":"a865d463086a2765995e7c23c3fee19b6839475e1cbfcd9f5aee911fdb3043e0"} Oct 02 13:47:40 crc kubenswrapper[4710]: I1002 13:47:40.914492 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerStarted","Data":"30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c"} Oct 02 13:47:42 crc kubenswrapper[4710]: I1002 13:47:42.928777 4710 generic.go:334] "Generic (PLEG): container finished" podID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerID="30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c" exitCode=0 Oct 02 13:47:42 crc kubenswrapper[4710]: I1002 13:47:42.930620 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerDied","Data":"30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c"} Oct 02 13:47:43 crc kubenswrapper[4710]: I1002 13:47:43.943730 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerStarted","Data":"0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb"} Oct 02 13:47:43 crc kubenswrapper[4710]: I1002 13:47:43.971923 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8mtxd" podStartSLOduration=2.559622148 podStartE2EDuration="5.971896197s" podCreationTimestamp="2025-10-02 13:47:38 +0000 UTC" firstStartedPulling="2025-10-02 13:47:39.892373639 +0000 UTC m=+3503.998784522" lastFinishedPulling="2025-10-02 13:47:43.304647688 +0000 UTC m=+3507.411058571" observedRunningTime="2025-10-02 13:47:43.968332817 +0000 UTC m=+3508.074743730" watchObservedRunningTime="2025-10-02 13:47:43.971896197 +0000 UTC m=+3508.078307090" Oct 02 13:47:48 crc kubenswrapper[4710]: I1002 13:47:48.812933 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:48 crc kubenswrapper[4710]: I1002 13:47:48.813564 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:49 crc kubenswrapper[4710]: I1002 13:47:49.863212 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8mtxd" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="registry-server" probeResult="failure" output=< Oct 02 13:47:49 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 13:47:49 crc kubenswrapper[4710]: > Oct 02 13:47:52 crc kubenswrapper[4710]: I1002 13:47:52.530611 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:47:52 crc kubenswrapper[4710]: I1002 13:47:52.531262 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:47:52 crc kubenswrapper[4710]: I1002 13:47:52.531392 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:47:52 crc kubenswrapper[4710]: I1002 13:47:52.532737 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ac006c31e441a8e868caaced27226ba6c8e96c7722cf7086f6c76445baa2a8f"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:47:52 crc kubenswrapper[4710]: I1002 13:47:52.532842 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://0ac006c31e441a8e868caaced27226ba6c8e96c7722cf7086f6c76445baa2a8f" gracePeriod=600 Oct 02 13:47:53 crc kubenswrapper[4710]: I1002 13:47:53.031461 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="0ac006c31e441a8e868caaced27226ba6c8e96c7722cf7086f6c76445baa2a8f" exitCode=0 Oct 02 13:47:53 crc kubenswrapper[4710]: I1002 13:47:53.031614 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"0ac006c31e441a8e868caaced27226ba6c8e96c7722cf7086f6c76445baa2a8f"} Oct 02 13:47:53 crc kubenswrapper[4710]: I1002 13:47:53.031825 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0"} Oct 02 13:47:53 crc kubenswrapper[4710]: I1002 13:47:53.031856 4710 scope.go:117] "RemoveContainer" containerID="f08bfdf0711c2e65441e01ba7b2b29eaa8e8b253093f020b5e968b69e7cf4572" Oct 02 13:47:58 crc kubenswrapper[4710]: I1002 13:47:58.867231 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:58 crc kubenswrapper[4710]: I1002 13:47:58.953455 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:47:59 crc kubenswrapper[4710]: I1002 13:47:59.106530 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8mtxd"] Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.106043 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8mtxd" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="registry-server" containerID="cri-o://0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb" gracePeriod=2 Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.581634 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.660582 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whh47\" (UniqueName: \"kubernetes.io/projected/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-kube-api-access-whh47\") pod \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.660628 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-catalog-content\") pod \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.678383 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-kube-api-access-whh47" (OuterVolumeSpecName: "kube-api-access-whh47") pod "7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" (UID: "7be1e2ee-6082-42d8-84e6-8709f9e8a8f3"). InnerVolumeSpecName "kube-api-access-whh47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.713134 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" (UID: "7be1e2ee-6082-42d8-84e6-8709f9e8a8f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.762301 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-utilities\") pod \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\" (UID: \"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3\") " Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.762973 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whh47\" (UniqueName: \"kubernetes.io/projected/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-kube-api-access-whh47\") on node \"crc\" DevicePath \"\"" Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.762997 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.763013 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-utilities" (OuterVolumeSpecName: "utilities") pod "7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" (UID: "7be1e2ee-6082-42d8-84e6-8709f9e8a8f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:48:00 crc kubenswrapper[4710]: I1002 13:48:00.864999 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.119735 4710 generic.go:334] "Generic (PLEG): container finished" podID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerID="0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb" exitCode=0 Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.119974 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerDied","Data":"0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb"} Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.120950 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8mtxd" event={"ID":"7be1e2ee-6082-42d8-84e6-8709f9e8a8f3","Type":"ContainerDied","Data":"a865d463086a2765995e7c23c3fee19b6839475e1cbfcd9f5aee911fdb3043e0"} Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.120017 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8mtxd" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.120983 4710 scope.go:117] "RemoveContainer" containerID="0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.159110 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8mtxd"] Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.180852 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8mtxd"] Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.181362 4710 scope.go:117] "RemoveContainer" containerID="30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.211791 4710 scope.go:117] "RemoveContainer" containerID="61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.275937 4710 scope.go:117] "RemoveContainer" containerID="0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb" Oct 02 13:48:01 crc kubenswrapper[4710]: E1002 13:48:01.276462 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb\": container with ID starting with 0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb not found: ID does not exist" containerID="0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.276500 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb"} err="failed to get container status \"0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb\": rpc error: code = NotFound desc = could not find container \"0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb\": container with ID starting with 0fed31c019b8f510097d71d9fd148326f1e47afdf41cf83decf4c221ac7c59bb not found: ID does not exist" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.276526 4710 scope.go:117] "RemoveContainer" containerID="30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c" Oct 02 13:48:01 crc kubenswrapper[4710]: E1002 13:48:01.276851 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c\": container with ID starting with 30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c not found: ID does not exist" containerID="30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.276882 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c"} err="failed to get container status \"30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c\": rpc error: code = NotFound desc = could not find container \"30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c\": container with ID starting with 30c506c00ece4f29ecbb823318fdda275a2b53157f16d3236de40f086930ca9c not found: ID does not exist" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.276900 4710 scope.go:117] "RemoveContainer" containerID="61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7" Oct 02 13:48:01 crc kubenswrapper[4710]: E1002 13:48:01.277179 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7\": container with ID starting with 61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7 not found: ID does not exist" containerID="61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7" Oct 02 13:48:01 crc kubenswrapper[4710]: I1002 13:48:01.277207 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7"} err="failed to get container status \"61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7\": rpc error: code = NotFound desc = could not find container \"61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7\": container with ID starting with 61cb6056e8df4a47b4696a76111ddc275a498da340e0496d395f95dbb56729c7 not found: ID does not exist" Oct 02 13:48:02 crc kubenswrapper[4710]: I1002 13:48:02.923474 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" path="/var/lib/kubelet/pods/7be1e2ee-6082-42d8-84e6-8709f9e8a8f3/volumes" Oct 02 13:48:29 crc kubenswrapper[4710]: I1002 13:48:29.822702 4710 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-76f964d7cf-dhr2b" podUID="407d1688-f160-4ee8-934d-9c76f353e4b9" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.088626 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2gc6d"] Oct 02 13:49:00 crc kubenswrapper[4710]: E1002 13:49:00.089770 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="extract-utilities" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.089789 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="extract-utilities" Oct 02 13:49:00 crc kubenswrapper[4710]: E1002 13:49:00.089840 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="registry-server" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.089850 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="registry-server" Oct 02 13:49:00 crc kubenswrapper[4710]: E1002 13:49:00.089866 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="extract-content" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.089873 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="extract-content" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.090199 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be1e2ee-6082-42d8-84e6-8709f9e8a8f3" containerName="registry-server" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.092238 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.106143 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gc6d"] Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.205657 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-utilities\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.205847 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czlfr\" (UniqueName: \"kubernetes.io/projected/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-kube-api-access-czlfr\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.206184 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-catalog-content\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.308269 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-catalog-content\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.308417 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-utilities\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.308463 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czlfr\" (UniqueName: \"kubernetes.io/projected/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-kube-api-access-czlfr\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.309231 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-catalog-content\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.309373 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-utilities\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.330140 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czlfr\" (UniqueName: \"kubernetes.io/projected/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-kube-api-access-czlfr\") pod \"community-operators-2gc6d\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:00 crc kubenswrapper[4710]: I1002 13:49:00.433607 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:01 crc kubenswrapper[4710]: I1002 13:49:01.015464 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gc6d"] Oct 02 13:49:01 crc kubenswrapper[4710]: I1002 13:49:01.719376 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerDied","Data":"0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9"} Oct 02 13:49:01 crc kubenswrapper[4710]: I1002 13:49:01.719215 4710 generic.go:334] "Generic (PLEG): container finished" podID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerID="0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9" exitCode=0 Oct 02 13:49:01 crc kubenswrapper[4710]: I1002 13:49:01.720355 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerStarted","Data":"18ed44e6b8073867d18c14c607558d2bfeceb7b691f2bc4a53f27e50c37efd26"} Oct 02 13:49:03 crc kubenswrapper[4710]: I1002 13:49:03.745528 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerStarted","Data":"ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c"} Oct 02 13:49:06 crc kubenswrapper[4710]: I1002 13:49:06.780328 4710 generic.go:334] "Generic (PLEG): container finished" podID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerID="ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c" exitCode=0 Oct 02 13:49:06 crc kubenswrapper[4710]: I1002 13:49:06.780402 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerDied","Data":"ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c"} Oct 02 13:49:08 crc kubenswrapper[4710]: I1002 13:49:08.800473 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerStarted","Data":"25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16"} Oct 02 13:49:08 crc kubenswrapper[4710]: I1002 13:49:08.825345 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2gc6d" podStartSLOduration=2.75608073 podStartE2EDuration="8.82532579s" podCreationTimestamp="2025-10-02 13:49:00 +0000 UTC" firstStartedPulling="2025-10-02 13:49:01.723877183 +0000 UTC m=+3585.830288066" lastFinishedPulling="2025-10-02 13:49:07.793122243 +0000 UTC m=+3591.899533126" observedRunningTime="2025-10-02 13:49:08.824666723 +0000 UTC m=+3592.931077616" watchObservedRunningTime="2025-10-02 13:49:08.82532579 +0000 UTC m=+3592.931736673" Oct 02 13:49:10 crc kubenswrapper[4710]: I1002 13:49:10.434445 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:10 crc kubenswrapper[4710]: I1002 13:49:10.435576 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:11 crc kubenswrapper[4710]: I1002 13:49:11.478557 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-2gc6d" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="registry-server" probeResult="failure" output=< Oct 02 13:49:11 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 13:49:11 crc kubenswrapper[4710]: > Oct 02 13:49:20 crc kubenswrapper[4710]: I1002 13:49:20.489334 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:20 crc kubenswrapper[4710]: I1002 13:49:20.561434 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:20 crc kubenswrapper[4710]: I1002 13:49:20.724907 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gc6d"] Oct 02 13:49:21 crc kubenswrapper[4710]: I1002 13:49:21.936605 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2gc6d" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="registry-server" containerID="cri-o://25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16" gracePeriod=2 Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.449771 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.476958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czlfr\" (UniqueName: \"kubernetes.io/projected/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-kube-api-access-czlfr\") pod \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.477049 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-catalog-content\") pod \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.477119 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-utilities\") pod \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\" (UID: \"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea\") " Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.478177 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-utilities" (OuterVolumeSpecName: "utilities") pod "90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" (UID: "90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.484955 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-kube-api-access-czlfr" (OuterVolumeSpecName: "kube-api-access-czlfr") pod "90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" (UID: "90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea"). InnerVolumeSpecName "kube-api-access-czlfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.535380 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" (UID: "90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.580434 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.580471 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.580483 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czlfr\" (UniqueName: \"kubernetes.io/projected/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea-kube-api-access-czlfr\") on node \"crc\" DevicePath \"\"" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.953856 4710 generic.go:334] "Generic (PLEG): container finished" podID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerID="25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16" exitCode=0 Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.953908 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gc6d" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.953904 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerDied","Data":"25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16"} Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.954041 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gc6d" event={"ID":"90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea","Type":"ContainerDied","Data":"18ed44e6b8073867d18c14c607558d2bfeceb7b691f2bc4a53f27e50c37efd26"} Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.954065 4710 scope.go:117] "RemoveContainer" containerID="25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.986835 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gc6d"] Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.991831 4710 scope.go:117] "RemoveContainer" containerID="ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c" Oct 02 13:49:22 crc kubenswrapper[4710]: I1002 13:49:22.996597 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2gc6d"] Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.023279 4710 scope.go:117] "RemoveContainer" containerID="0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9" Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.072607 4710 scope.go:117] "RemoveContainer" containerID="25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16" Oct 02 13:49:23 crc kubenswrapper[4710]: E1002 13:49:23.073177 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16\": container with ID starting with 25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16 not found: ID does not exist" containerID="25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16" Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.073213 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16"} err="failed to get container status \"25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16\": rpc error: code = NotFound desc = could not find container \"25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16\": container with ID starting with 25765ae9009b3987c8e5f21192aeaf55e81bc51b483ae727d14504b56adfcf16 not found: ID does not exist" Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.073241 4710 scope.go:117] "RemoveContainer" containerID="ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c" Oct 02 13:49:23 crc kubenswrapper[4710]: E1002 13:49:23.073516 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c\": container with ID starting with ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c not found: ID does not exist" containerID="ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c" Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.073545 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c"} err="failed to get container status \"ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c\": rpc error: code = NotFound desc = could not find container \"ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c\": container with ID starting with ef9edb3b9fd8a85561589f074436d5839558bfc963c0a9671027e3d0d784c14c not found: ID does not exist" Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.073564 4710 scope.go:117] "RemoveContainer" containerID="0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9" Oct 02 13:49:23 crc kubenswrapper[4710]: E1002 13:49:23.074045 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9\": container with ID starting with 0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9 not found: ID does not exist" containerID="0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9" Oct 02 13:49:23 crc kubenswrapper[4710]: I1002 13:49:23.074071 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9"} err="failed to get container status \"0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9\": rpc error: code = NotFound desc = could not find container \"0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9\": container with ID starting with 0e8174284dc4f4540ba070138a7f412424fd164e823dc5ab2e978859fa0cc6b9 not found: ID does not exist" Oct 02 13:49:24 crc kubenswrapper[4710]: I1002 13:49:24.919138 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" path="/var/lib/kubelet/pods/90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea/volumes" Oct 02 13:49:52 crc kubenswrapper[4710]: I1002 13:49:52.529890 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:49:52 crc kubenswrapper[4710]: I1002 13:49:52.530309 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:50:22 crc kubenswrapper[4710]: I1002 13:50:22.529941 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:50:22 crc kubenswrapper[4710]: I1002 13:50:22.530475 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.530658 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.531126 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.531174 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.531941 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.531991 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" gracePeriod=600 Oct 02 13:50:52 crc kubenswrapper[4710]: E1002 13:50:52.652581 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.865718 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" exitCode=0 Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.865779 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0"} Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.865847 4710 scope.go:117] "RemoveContainer" containerID="0ac006c31e441a8e868caaced27226ba6c8e96c7722cf7086f6c76445baa2a8f" Oct 02 13:50:52 crc kubenswrapper[4710]: I1002 13:50:52.866499 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:50:52 crc kubenswrapper[4710]: E1002 13:50:52.866829 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:51:06 crc kubenswrapper[4710]: I1002 13:51:06.911287 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:51:06 crc kubenswrapper[4710]: E1002 13:51:06.912029 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:51:21 crc kubenswrapper[4710]: I1002 13:51:21.904492 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:51:21 crc kubenswrapper[4710]: E1002 13:51:21.905491 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:51:34 crc kubenswrapper[4710]: I1002 13:51:34.904923 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:51:34 crc kubenswrapper[4710]: E1002 13:51:34.905977 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:51:45 crc kubenswrapper[4710]: I1002 13:51:45.905186 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:51:45 crc kubenswrapper[4710]: E1002 13:51:45.905963 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:51:58 crc kubenswrapper[4710]: I1002 13:51:58.904025 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:51:58 crc kubenswrapper[4710]: E1002 13:51:58.904858 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:52:10 crc kubenswrapper[4710]: I1002 13:52:10.905324 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:52:10 crc kubenswrapper[4710]: E1002 13:52:10.906048 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:52:24 crc kubenswrapper[4710]: I1002 13:52:24.904496 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:52:24 crc kubenswrapper[4710]: E1002 13:52:24.905545 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:52:31 crc kubenswrapper[4710]: E1002 13:52:31.362791 4710 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.103:54300->38.102.83.103:33121: write tcp 38.102.83.103:54300->38.102.83.103:33121: write: broken pipe Oct 02 13:52:38 crc kubenswrapper[4710]: I1002 13:52:38.904737 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:52:38 crc kubenswrapper[4710]: E1002 13:52:38.905499 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:52:49 crc kubenswrapper[4710]: I1002 13:52:49.904350 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:52:49 crc kubenswrapper[4710]: E1002 13:52:49.904932 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:53:01 crc kubenswrapper[4710]: I1002 13:53:01.905317 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:53:01 crc kubenswrapper[4710]: E1002 13:53:01.906200 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:53:16 crc kubenswrapper[4710]: I1002 13:53:16.912931 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:53:16 crc kubenswrapper[4710]: E1002 13:53:16.913906 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:53:27 crc kubenswrapper[4710]: I1002 13:53:27.905225 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:53:27 crc kubenswrapper[4710]: E1002 13:53:27.906082 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:53:42 crc kubenswrapper[4710]: I1002 13:53:42.904625 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:53:42 crc kubenswrapper[4710]: E1002 13:53:42.905391 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:53:57 crc kubenswrapper[4710]: I1002 13:53:57.905618 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:53:57 crc kubenswrapper[4710]: E1002 13:53:57.906644 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:54:09 crc kubenswrapper[4710]: I1002 13:54:09.905142 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:54:09 crc kubenswrapper[4710]: E1002 13:54:09.905797 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:54:24 crc kubenswrapper[4710]: I1002 13:54:24.905773 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:54:24 crc kubenswrapper[4710]: E1002 13:54:24.907023 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:54:36 crc kubenswrapper[4710]: I1002 13:54:36.912592 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:54:36 crc kubenswrapper[4710]: E1002 13:54:36.923102 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.834236 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4q4cl"] Oct 02 13:54:47 crc kubenswrapper[4710]: E1002 13:54:47.835252 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="registry-server" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.835269 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="registry-server" Oct 02 13:54:47 crc kubenswrapper[4710]: E1002 13:54:47.835293 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="extract-utilities" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.835301 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="extract-utilities" Oct 02 13:54:47 crc kubenswrapper[4710]: E1002 13:54:47.835340 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="extract-content" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.835348 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="extract-content" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.835632 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="90bb0888-7ae4-4f3f-bafe-b79d2ec8aeea" containerName="registry-server" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.837771 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.844863 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4q4cl"] Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.904601 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:54:47 crc kubenswrapper[4710]: E1002 13:54:47.905010 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.981981 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-utilities\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.982093 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5njb\" (UniqueName: \"kubernetes.io/projected/7175321d-da45-445e-8c4f-3bd1b5daa70c-kube-api-access-r5njb\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:47 crc kubenswrapper[4710]: I1002 13:54:47.982165 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-catalog-content\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.084115 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-utilities\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.084531 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5njb\" (UniqueName: \"kubernetes.io/projected/7175321d-da45-445e-8c4f-3bd1b5daa70c-kube-api-access-r5njb\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.084595 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-catalog-content\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.084621 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-utilities\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.084932 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-catalog-content\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.102597 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5njb\" (UniqueName: \"kubernetes.io/projected/7175321d-da45-445e-8c4f-3bd1b5daa70c-kube-api-access-r5njb\") pod \"redhat-operators-4q4cl\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.162250 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:48 crc kubenswrapper[4710]: I1002 13:54:48.642174 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4q4cl"] Oct 02 13:54:49 crc kubenswrapper[4710]: I1002 13:54:49.553399 4710 generic.go:334] "Generic (PLEG): container finished" podID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerID="cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12" exitCode=0 Oct 02 13:54:49 crc kubenswrapper[4710]: I1002 13:54:49.553620 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerDied","Data":"cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12"} Oct 02 13:54:49 crc kubenswrapper[4710]: I1002 13:54:49.553891 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerStarted","Data":"46e152352472227abdb68e196ea56fd9274e9d0d0be55ddcb7e928031e22d0dc"} Oct 02 13:54:49 crc kubenswrapper[4710]: I1002 13:54:49.566764 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 13:54:51 crc kubenswrapper[4710]: I1002 13:54:51.582309 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerStarted","Data":"e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea"} Oct 02 13:54:52 crc kubenswrapper[4710]: I1002 13:54:52.594480 4710 generic.go:334] "Generic (PLEG): container finished" podID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerID="e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea" exitCode=0 Oct 02 13:54:52 crc kubenswrapper[4710]: I1002 13:54:52.594546 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerDied","Data":"e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea"} Oct 02 13:54:53 crc kubenswrapper[4710]: I1002 13:54:53.608696 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerStarted","Data":"6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6"} Oct 02 13:54:53 crc kubenswrapper[4710]: I1002 13:54:53.639632 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4q4cl" podStartSLOduration=2.964461234 podStartE2EDuration="6.639603563s" podCreationTimestamp="2025-10-02 13:54:47 +0000 UTC" firstStartedPulling="2025-10-02 13:54:49.56651761 +0000 UTC m=+3933.672928493" lastFinishedPulling="2025-10-02 13:54:53.241659889 +0000 UTC m=+3937.348070822" observedRunningTime="2025-10-02 13:54:53.627536098 +0000 UTC m=+3937.733946991" watchObservedRunningTime="2025-10-02 13:54:53.639603563 +0000 UTC m=+3937.746014456" Oct 02 13:54:58 crc kubenswrapper[4710]: I1002 13:54:58.163122 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:58 crc kubenswrapper[4710]: I1002 13:54:58.163904 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:58 crc kubenswrapper[4710]: I1002 13:54:58.221801 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:58 crc kubenswrapper[4710]: I1002 13:54:58.721912 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:54:58 crc kubenswrapper[4710]: I1002 13:54:58.783904 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4q4cl"] Oct 02 13:55:00 crc kubenswrapper[4710]: I1002 13:55:00.686831 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4q4cl" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="registry-server" containerID="cri-o://6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6" gracePeriod=2 Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.163064 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.259805 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5njb\" (UniqueName: \"kubernetes.io/projected/7175321d-da45-445e-8c4f-3bd1b5daa70c-kube-api-access-r5njb\") pod \"7175321d-da45-445e-8c4f-3bd1b5daa70c\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.259873 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-catalog-content\") pod \"7175321d-da45-445e-8c4f-3bd1b5daa70c\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.260027 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-utilities\") pod \"7175321d-da45-445e-8c4f-3bd1b5daa70c\" (UID: \"7175321d-da45-445e-8c4f-3bd1b5daa70c\") " Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.261022 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-utilities" (OuterVolumeSpecName: "utilities") pod "7175321d-da45-445e-8c4f-3bd1b5daa70c" (UID: "7175321d-da45-445e-8c4f-3bd1b5daa70c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.266965 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7175321d-da45-445e-8c4f-3bd1b5daa70c-kube-api-access-r5njb" (OuterVolumeSpecName: "kube-api-access-r5njb") pod "7175321d-da45-445e-8c4f-3bd1b5daa70c" (UID: "7175321d-da45-445e-8c4f-3bd1b5daa70c"). InnerVolumeSpecName "kube-api-access-r5njb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.362555 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.362584 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5njb\" (UniqueName: \"kubernetes.io/projected/7175321d-da45-445e-8c4f-3bd1b5daa70c-kube-api-access-r5njb\") on node \"crc\" DevicePath \"\"" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.701650 4710 generic.go:334] "Generic (PLEG): container finished" podID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerID="6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6" exitCode=0 Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.701703 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerDied","Data":"6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6"} Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.701723 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4q4cl" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.701766 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4q4cl" event={"ID":"7175321d-da45-445e-8c4f-3bd1b5daa70c","Type":"ContainerDied","Data":"46e152352472227abdb68e196ea56fd9274e9d0d0be55ddcb7e928031e22d0dc"} Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.701790 4710 scope.go:117] "RemoveContainer" containerID="6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.753777 4710 scope.go:117] "RemoveContainer" containerID="e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.783903 4710 scope.go:117] "RemoveContainer" containerID="cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.870451 4710 scope.go:117] "RemoveContainer" containerID="6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6" Oct 02 13:55:01 crc kubenswrapper[4710]: E1002 13:55:01.871113 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6\": container with ID starting with 6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6 not found: ID does not exist" containerID="6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.871141 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6"} err="failed to get container status \"6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6\": rpc error: code = NotFound desc = could not find container \"6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6\": container with ID starting with 6cc26becfb374c2701683c75e96b983650677c7bfdd70ea628f449b0d40a51b6 not found: ID does not exist" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.871162 4710 scope.go:117] "RemoveContainer" containerID="e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea" Oct 02 13:55:01 crc kubenswrapper[4710]: E1002 13:55:01.872259 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea\": container with ID starting with e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea not found: ID does not exist" containerID="e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.872314 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea"} err="failed to get container status \"e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea\": rpc error: code = NotFound desc = could not find container \"e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea\": container with ID starting with e94b8967962ee451b55d020b8a6fb97c75f3451d588680b376f6425677a7ffea not found: ID does not exist" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.872349 4710 scope.go:117] "RemoveContainer" containerID="cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12" Oct 02 13:55:01 crc kubenswrapper[4710]: E1002 13:55:01.872996 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12\": container with ID starting with cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12 not found: ID does not exist" containerID="cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12" Oct 02 13:55:01 crc kubenswrapper[4710]: I1002 13:55:01.873022 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12"} err="failed to get container status \"cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12\": rpc error: code = NotFound desc = could not find container \"cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12\": container with ID starting with cf9718ecf641f7528473f1105e8039d84477ce6e43ef8043d4dde2cb8f61af12 not found: ID does not exist" Oct 02 13:55:02 crc kubenswrapper[4710]: I1002 13:55:02.766115 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7175321d-da45-445e-8c4f-3bd1b5daa70c" (UID: "7175321d-da45-445e-8c4f-3bd1b5daa70c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:55:02 crc kubenswrapper[4710]: I1002 13:55:02.790613 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7175321d-da45-445e-8c4f-3bd1b5daa70c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:55:02 crc kubenswrapper[4710]: I1002 13:55:02.905681 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:55:02 crc kubenswrapper[4710]: E1002 13:55:02.906209 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:55:02 crc kubenswrapper[4710]: I1002 13:55:02.944041 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4q4cl"] Oct 02 13:55:02 crc kubenswrapper[4710]: I1002 13:55:02.959685 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4q4cl"] Oct 02 13:55:04 crc kubenswrapper[4710]: I1002 13:55:04.917182 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" path="/var/lib/kubelet/pods/7175321d-da45-445e-8c4f-3bd1b5daa70c/volumes" Oct 02 13:55:15 crc kubenswrapper[4710]: I1002 13:55:15.904418 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:55:15 crc kubenswrapper[4710]: E1002 13:55:15.905542 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.863444 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4kt2v"] Oct 02 13:55:24 crc kubenswrapper[4710]: E1002 13:55:24.864539 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="extract-utilities" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.864584 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="extract-utilities" Oct 02 13:55:24 crc kubenswrapper[4710]: E1002 13:55:24.864600 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="registry-server" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.864609 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="registry-server" Oct 02 13:55:24 crc kubenswrapper[4710]: E1002 13:55:24.864623 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="extract-content" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.864632 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="extract-content" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.865108 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="7175321d-da45-445e-8c4f-3bd1b5daa70c" containerName="registry-server" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.868668 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:24 crc kubenswrapper[4710]: I1002 13:55:24.876331 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kt2v"] Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.041821 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-catalog-content\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.041883 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w26wk\" (UniqueName: \"kubernetes.io/projected/c1d10c5e-be18-4906-82dd-88a27e185efe-kube-api-access-w26wk\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.041922 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-utilities\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.143931 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-utilities\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.144502 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-catalog-content\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.144665 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w26wk\" (UniqueName: \"kubernetes.io/projected/c1d10c5e-be18-4906-82dd-88a27e185efe-kube-api-access-w26wk\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.144795 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-utilities\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.144953 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-catalog-content\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.166729 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w26wk\" (UniqueName: \"kubernetes.io/projected/c1d10c5e-be18-4906-82dd-88a27e185efe-kube-api-access-w26wk\") pod \"redhat-marketplace-4kt2v\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.195976 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.664473 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kt2v"] Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.972260 4710 generic.go:334] "Generic (PLEG): container finished" podID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerID="93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028" exitCode=0 Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.972412 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kt2v" event={"ID":"c1d10c5e-be18-4906-82dd-88a27e185efe","Type":"ContainerDied","Data":"93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028"} Oct 02 13:55:25 crc kubenswrapper[4710]: I1002 13:55:25.972608 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kt2v" event={"ID":"c1d10c5e-be18-4906-82dd-88a27e185efe","Type":"ContainerStarted","Data":"948a651b522eb9fe039a72628ca94b67c37644cf5e5a1eabefd66569b920546a"} Oct 02 13:55:27 crc kubenswrapper[4710]: I1002 13:55:27.992089 4710 generic.go:334] "Generic (PLEG): container finished" podID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerID="c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287" exitCode=0 Oct 02 13:55:27 crc kubenswrapper[4710]: I1002 13:55:27.992148 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kt2v" event={"ID":"c1d10c5e-be18-4906-82dd-88a27e185efe","Type":"ContainerDied","Data":"c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287"} Oct 02 13:55:28 crc kubenswrapper[4710]: I1002 13:55:28.906167 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:55:28 crc kubenswrapper[4710]: E1002 13:55:28.907204 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:55:29 crc kubenswrapper[4710]: I1002 13:55:29.003207 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kt2v" event={"ID":"c1d10c5e-be18-4906-82dd-88a27e185efe","Type":"ContainerStarted","Data":"9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df"} Oct 02 13:55:29 crc kubenswrapper[4710]: I1002 13:55:29.029034 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4kt2v" podStartSLOduration=2.487586961 podStartE2EDuration="5.029008865s" podCreationTimestamp="2025-10-02 13:55:24 +0000 UTC" firstStartedPulling="2025-10-02 13:55:25.97484927 +0000 UTC m=+3970.081260153" lastFinishedPulling="2025-10-02 13:55:28.516271164 +0000 UTC m=+3972.622682057" observedRunningTime="2025-10-02 13:55:29.028984504 +0000 UTC m=+3973.135395407" watchObservedRunningTime="2025-10-02 13:55:29.029008865 +0000 UTC m=+3973.135419748" Oct 02 13:55:35 crc kubenswrapper[4710]: I1002 13:55:35.197240 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:35 crc kubenswrapper[4710]: I1002 13:55:35.199169 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:35 crc kubenswrapper[4710]: I1002 13:55:35.291646 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:36 crc kubenswrapper[4710]: I1002 13:55:36.188985 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:36 crc kubenswrapper[4710]: I1002 13:55:36.258722 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kt2v"] Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.125583 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4kt2v" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="registry-server" containerID="cri-o://9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df" gracePeriod=2 Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.614903 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.643979 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-catalog-content\") pod \"c1d10c5e-be18-4906-82dd-88a27e185efe\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.644052 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w26wk\" (UniqueName: \"kubernetes.io/projected/c1d10c5e-be18-4906-82dd-88a27e185efe-kube-api-access-w26wk\") pod \"c1d10c5e-be18-4906-82dd-88a27e185efe\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.644179 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-utilities\") pod \"c1d10c5e-be18-4906-82dd-88a27e185efe\" (UID: \"c1d10c5e-be18-4906-82dd-88a27e185efe\") " Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.646620 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-utilities" (OuterVolumeSpecName: "utilities") pod "c1d10c5e-be18-4906-82dd-88a27e185efe" (UID: "c1d10c5e-be18-4906-82dd-88a27e185efe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.653867 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1d10c5e-be18-4906-82dd-88a27e185efe-kube-api-access-w26wk" (OuterVolumeSpecName: "kube-api-access-w26wk") pod "c1d10c5e-be18-4906-82dd-88a27e185efe" (UID: "c1d10c5e-be18-4906-82dd-88a27e185efe"). InnerVolumeSpecName "kube-api-access-w26wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.675138 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1d10c5e-be18-4906-82dd-88a27e185efe" (UID: "c1d10c5e-be18-4906-82dd-88a27e185efe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.745403 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.745436 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d10c5e-be18-4906-82dd-88a27e185efe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:55:38 crc kubenswrapper[4710]: I1002 13:55:38.745446 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w26wk\" (UniqueName: \"kubernetes.io/projected/c1d10c5e-be18-4906-82dd-88a27e185efe-kube-api-access-w26wk\") on node \"crc\" DevicePath \"\"" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.139651 4710 generic.go:334] "Generic (PLEG): container finished" podID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerID="9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df" exitCode=0 Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.139904 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kt2v" event={"ID":"c1d10c5e-be18-4906-82dd-88a27e185efe","Type":"ContainerDied","Data":"9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df"} Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.139998 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kt2v" event={"ID":"c1d10c5e-be18-4906-82dd-88a27e185efe","Type":"ContainerDied","Data":"948a651b522eb9fe039a72628ca94b67c37644cf5e5a1eabefd66569b920546a"} Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.139951 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kt2v" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.140042 4710 scope.go:117] "RemoveContainer" containerID="9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.166782 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kt2v"] Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.173216 4710 scope.go:117] "RemoveContainer" containerID="c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.175731 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kt2v"] Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.202861 4710 scope.go:117] "RemoveContainer" containerID="93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.275674 4710 scope.go:117] "RemoveContainer" containerID="9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df" Oct 02 13:55:39 crc kubenswrapper[4710]: E1002 13:55:39.276257 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df\": container with ID starting with 9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df not found: ID does not exist" containerID="9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.276284 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df"} err="failed to get container status \"9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df\": rpc error: code = NotFound desc = could not find container \"9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df\": container with ID starting with 9397b2fd95a25079cd1dd6d6ef67ed6f3155d7f9c274d484ce328b9bc67820df not found: ID does not exist" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.276303 4710 scope.go:117] "RemoveContainer" containerID="c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287" Oct 02 13:55:39 crc kubenswrapper[4710]: E1002 13:55:39.276562 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287\": container with ID starting with c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287 not found: ID does not exist" containerID="c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.276595 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287"} err="failed to get container status \"c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287\": rpc error: code = NotFound desc = could not find container \"c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287\": container with ID starting with c96abe143a1ebc8dfaae7f1319e78f945fa8eb8e32e53394467f7b25f9c23287 not found: ID does not exist" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.276609 4710 scope.go:117] "RemoveContainer" containerID="93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028" Oct 02 13:55:39 crc kubenswrapper[4710]: E1002 13:55:39.276951 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028\": container with ID starting with 93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028 not found: ID does not exist" containerID="93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028" Oct 02 13:55:39 crc kubenswrapper[4710]: I1002 13:55:39.276973 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028"} err="failed to get container status \"93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028\": rpc error: code = NotFound desc = could not find container \"93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028\": container with ID starting with 93a952ed0613abb950fd8e6568ffcea4cde69fc775e1a38119652ea2517c8028 not found: ID does not exist" Oct 02 13:55:40 crc kubenswrapper[4710]: I1002 13:55:40.926011 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" path="/var/lib/kubelet/pods/c1d10c5e-be18-4906-82dd-88a27e185efe/volumes" Oct 02 13:55:41 crc kubenswrapper[4710]: I1002 13:55:41.904930 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:55:41 crc kubenswrapper[4710]: E1002 13:55:41.905293 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 13:55:53 crc kubenswrapper[4710]: I1002 13:55:53.904884 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:55:54 crc kubenswrapper[4710]: I1002 13:55:54.314926 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"357899fea19c2894c442c75e131da52c0455c786dbcf957b2a9376ad405bb029"} Oct 02 13:58:22 crc kubenswrapper[4710]: I1002 13:58:22.530847 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:58:22 crc kubenswrapper[4710]: I1002 13:58:22.531408 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.809829 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pr2w8"] Oct 02 13:58:23 crc kubenswrapper[4710]: E1002 13:58:23.810556 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="extract-content" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.810569 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="extract-content" Oct 02 13:58:23 crc kubenswrapper[4710]: E1002 13:58:23.810602 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="registry-server" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.810608 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="registry-server" Oct 02 13:58:23 crc kubenswrapper[4710]: E1002 13:58:23.810625 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="extract-utilities" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.810631 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="extract-utilities" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.810835 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d10c5e-be18-4906-82dd-88a27e185efe" containerName="registry-server" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.812477 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.843136 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pr2w8"] Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.975290 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmmwx\" (UniqueName: \"kubernetes.io/projected/a15da537-7420-4469-9d62-7301e8d99897-kube-api-access-pmmwx\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.975685 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-catalog-content\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:23 crc kubenswrapper[4710]: I1002 13:58:23.976016 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-utilities\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.077975 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-catalog-content\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.078147 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-utilities\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.078194 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmmwx\" (UniqueName: \"kubernetes.io/projected/a15da537-7420-4469-9d62-7301e8d99897-kube-api-access-pmmwx\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.078610 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-catalog-content\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.078680 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-utilities\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.098288 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmmwx\" (UniqueName: \"kubernetes.io/projected/a15da537-7420-4469-9d62-7301e8d99897-kube-api-access-pmmwx\") pod \"certified-operators-pr2w8\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.145365 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.667254 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pr2w8"] Oct 02 13:58:24 crc kubenswrapper[4710]: I1002 13:58:24.916370 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerStarted","Data":"d50cdc9b8c6fa14e041cd8d3427416ecdc1a859bdbda93899676492bdf49a591"} Oct 02 13:58:25 crc kubenswrapper[4710]: I1002 13:58:25.930141 4710 generic.go:334] "Generic (PLEG): container finished" podID="a15da537-7420-4469-9d62-7301e8d99897" containerID="892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19" exitCode=0 Oct 02 13:58:25 crc kubenswrapper[4710]: I1002 13:58:25.930209 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerDied","Data":"892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19"} Oct 02 13:58:27 crc kubenswrapper[4710]: I1002 13:58:27.955568 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerStarted","Data":"9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2"} Oct 02 13:58:28 crc kubenswrapper[4710]: I1002 13:58:28.966999 4710 generic.go:334] "Generic (PLEG): container finished" podID="a15da537-7420-4469-9d62-7301e8d99897" containerID="9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2" exitCode=0 Oct 02 13:58:28 crc kubenswrapper[4710]: I1002 13:58:28.967061 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerDied","Data":"9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2"} Oct 02 13:58:30 crc kubenswrapper[4710]: I1002 13:58:30.990832 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerStarted","Data":"79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b"} Oct 02 13:58:31 crc kubenswrapper[4710]: I1002 13:58:31.018838 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pr2w8" podStartSLOduration=4.244044354 podStartE2EDuration="8.018812186s" podCreationTimestamp="2025-10-02 13:58:23 +0000 UTC" firstStartedPulling="2025-10-02 13:58:25.933190069 +0000 UTC m=+4150.039600952" lastFinishedPulling="2025-10-02 13:58:29.707957901 +0000 UTC m=+4153.814368784" observedRunningTime="2025-10-02 13:58:31.015089312 +0000 UTC m=+4155.121500235" watchObservedRunningTime="2025-10-02 13:58:31.018812186 +0000 UTC m=+4155.125223089" Oct 02 13:58:34 crc kubenswrapper[4710]: I1002 13:58:34.146101 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:34 crc kubenswrapper[4710]: I1002 13:58:34.148343 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:34 crc kubenswrapper[4710]: I1002 13:58:34.235577 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:35 crc kubenswrapper[4710]: I1002 13:58:35.113307 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:35 crc kubenswrapper[4710]: I1002 13:58:35.174575 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pr2w8"] Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.062941 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pr2w8" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="registry-server" containerID="cri-o://79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b" gracePeriod=2 Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.552095 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.705125 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-catalog-content\") pod \"a15da537-7420-4469-9d62-7301e8d99897\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.705191 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-utilities\") pod \"a15da537-7420-4469-9d62-7301e8d99897\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.705311 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmmwx\" (UniqueName: \"kubernetes.io/projected/a15da537-7420-4469-9d62-7301e8d99897-kube-api-access-pmmwx\") pod \"a15da537-7420-4469-9d62-7301e8d99897\" (UID: \"a15da537-7420-4469-9d62-7301e8d99897\") " Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.706153 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-utilities" (OuterVolumeSpecName: "utilities") pod "a15da537-7420-4469-9d62-7301e8d99897" (UID: "a15da537-7420-4469-9d62-7301e8d99897"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.706374 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.714013 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a15da537-7420-4469-9d62-7301e8d99897-kube-api-access-pmmwx" (OuterVolumeSpecName: "kube-api-access-pmmwx") pod "a15da537-7420-4469-9d62-7301e8d99897" (UID: "a15da537-7420-4469-9d62-7301e8d99897"). InnerVolumeSpecName "kube-api-access-pmmwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.753761 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a15da537-7420-4469-9d62-7301e8d99897" (UID: "a15da537-7420-4469-9d62-7301e8d99897"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.808440 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15da537-7420-4469-9d62-7301e8d99897-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 13:58:37 crc kubenswrapper[4710]: I1002 13:58:37.808481 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmmwx\" (UniqueName: \"kubernetes.io/projected/a15da537-7420-4469-9d62-7301e8d99897-kube-api-access-pmmwx\") on node \"crc\" DevicePath \"\"" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.079271 4710 generic.go:334] "Generic (PLEG): container finished" podID="a15da537-7420-4469-9d62-7301e8d99897" containerID="79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b" exitCode=0 Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.079332 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerDied","Data":"79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b"} Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.079359 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr2w8" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.079376 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr2w8" event={"ID":"a15da537-7420-4469-9d62-7301e8d99897","Type":"ContainerDied","Data":"d50cdc9b8c6fa14e041cd8d3427416ecdc1a859bdbda93899676492bdf49a591"} Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.079393 4710 scope.go:117] "RemoveContainer" containerID="79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.113875 4710 scope.go:117] "RemoveContainer" containerID="9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.144651 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pr2w8"] Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.151472 4710 scope.go:117] "RemoveContainer" containerID="892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.159089 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pr2w8"] Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.216128 4710 scope.go:117] "RemoveContainer" containerID="79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b" Oct 02 13:58:38 crc kubenswrapper[4710]: E1002 13:58:38.216845 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b\": container with ID starting with 79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b not found: ID does not exist" containerID="79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.216878 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b"} err="failed to get container status \"79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b\": rpc error: code = NotFound desc = could not find container \"79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b\": container with ID starting with 79c7210c81cdae728390f13922c843175aee9d2d887432d91e829fca6809ca8b not found: ID does not exist" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.216900 4710 scope.go:117] "RemoveContainer" containerID="9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2" Oct 02 13:58:38 crc kubenswrapper[4710]: E1002 13:58:38.217103 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2\": container with ID starting with 9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2 not found: ID does not exist" containerID="9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.217121 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2"} err="failed to get container status \"9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2\": rpc error: code = NotFound desc = could not find container \"9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2\": container with ID starting with 9612471e62c2b79b4b246ab3f5b5a93ea52620c4c4158b22afef64411ffe19a2 not found: ID does not exist" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.217133 4710 scope.go:117] "RemoveContainer" containerID="892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19" Oct 02 13:58:38 crc kubenswrapper[4710]: E1002 13:58:38.217281 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19\": container with ID starting with 892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19 not found: ID does not exist" containerID="892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.217296 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19"} err="failed to get container status \"892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19\": rpc error: code = NotFound desc = could not find container \"892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19\": container with ID starting with 892ee3ea86e478b502ce1dd7aed739bafcc323fd7e3ae125243097fd228e6a19 not found: ID does not exist" Oct 02 13:58:38 crc kubenswrapper[4710]: I1002 13:58:38.927067 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a15da537-7420-4469-9d62-7301e8d99897" path="/var/lib/kubelet/pods/a15da537-7420-4469-9d62-7301e8d99897/volumes" Oct 02 13:58:52 crc kubenswrapper[4710]: I1002 13:58:52.530302 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:58:52 crc kubenswrapper[4710]: I1002 13:58:52.530683 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:59:22 crc kubenswrapper[4710]: I1002 13:59:22.530174 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 13:59:22 crc kubenswrapper[4710]: I1002 13:59:22.530897 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 13:59:22 crc kubenswrapper[4710]: I1002 13:59:22.530960 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 13:59:22 crc kubenswrapper[4710]: I1002 13:59:22.531952 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"357899fea19c2894c442c75e131da52c0455c786dbcf957b2a9376ad405bb029"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 13:59:22 crc kubenswrapper[4710]: I1002 13:59:22.532062 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://357899fea19c2894c442c75e131da52c0455c786dbcf957b2a9376ad405bb029" gracePeriod=600 Oct 02 13:59:23 crc kubenswrapper[4710]: I1002 13:59:23.604128 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="357899fea19c2894c442c75e131da52c0455c786dbcf957b2a9376ad405bb029" exitCode=0 Oct 02 13:59:23 crc kubenswrapper[4710]: I1002 13:59:23.604420 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"357899fea19c2894c442c75e131da52c0455c786dbcf957b2a9376ad405bb029"} Oct 02 13:59:23 crc kubenswrapper[4710]: I1002 13:59:23.604450 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b"} Oct 02 13:59:23 crc kubenswrapper[4710]: I1002 13:59:23.604471 4710 scope.go:117] "RemoveContainer" containerID="177a92e5e97049190b8f95a98280f7833de7b86a9fef70df2f9763934e5dbaa0" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.215358 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q646l"] Oct 02 13:59:41 crc kubenswrapper[4710]: E1002 13:59:41.216419 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="extract-utilities" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.216437 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="extract-utilities" Oct 02 13:59:41 crc kubenswrapper[4710]: E1002 13:59:41.216466 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="registry-server" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.216475 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="registry-server" Oct 02 13:59:41 crc kubenswrapper[4710]: E1002 13:59:41.216520 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="extract-content" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.216530 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="extract-content" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.216830 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a15da537-7420-4469-9d62-7301e8d99897" containerName="registry-server" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.221906 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.239369 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-utilities\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.239428 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-catalog-content\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.239487 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbqp5\" (UniqueName: \"kubernetes.io/projected/18853a2b-28e7-45b8-b0f8-92343a7035eb-kube-api-access-vbqp5\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.247344 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q646l"] Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.341786 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-utilities\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.341845 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-catalog-content\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.341908 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbqp5\" (UniqueName: \"kubernetes.io/projected/18853a2b-28e7-45b8-b0f8-92343a7035eb-kube-api-access-vbqp5\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.342386 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-utilities\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.342400 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-catalog-content\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.360422 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbqp5\" (UniqueName: \"kubernetes.io/projected/18853a2b-28e7-45b8-b0f8-92343a7035eb-kube-api-access-vbqp5\") pod \"community-operators-q646l\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:41 crc kubenswrapper[4710]: I1002 13:59:41.547814 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:42 crc kubenswrapper[4710]: I1002 13:59:42.079684 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q646l"] Oct 02 13:59:42 crc kubenswrapper[4710]: I1002 13:59:42.814417 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerStarted","Data":"d26f7480d4b19929b91fa15ac92b6fe850dedd7248463f3f8bb8ac3eb95dbe06"} Oct 02 13:59:43 crc kubenswrapper[4710]: I1002 13:59:43.828358 4710 generic.go:334] "Generic (PLEG): container finished" podID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerID="f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215" exitCode=0 Oct 02 13:59:43 crc kubenswrapper[4710]: I1002 13:59:43.828652 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerDied","Data":"f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215"} Oct 02 13:59:45 crc kubenswrapper[4710]: I1002 13:59:45.851312 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerStarted","Data":"19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8"} Oct 02 13:59:47 crc kubenswrapper[4710]: I1002 13:59:47.873782 4710 generic.go:334] "Generic (PLEG): container finished" podID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerID="19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8" exitCode=0 Oct 02 13:59:47 crc kubenswrapper[4710]: I1002 13:59:47.873890 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerDied","Data":"19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8"} Oct 02 13:59:48 crc kubenswrapper[4710]: I1002 13:59:48.916513 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerStarted","Data":"a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59"} Oct 02 13:59:48 crc kubenswrapper[4710]: I1002 13:59:48.947234 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q646l" podStartSLOduration=3.454233109 podStartE2EDuration="7.947207017s" podCreationTimestamp="2025-10-02 13:59:41 +0000 UTC" firstStartedPulling="2025-10-02 13:59:43.830360923 +0000 UTC m=+4227.936771826" lastFinishedPulling="2025-10-02 13:59:48.323334851 +0000 UTC m=+4232.429745734" observedRunningTime="2025-10-02 13:59:48.941028591 +0000 UTC m=+4233.047439474" watchObservedRunningTime="2025-10-02 13:59:48.947207017 +0000 UTC m=+4233.053617900" Oct 02 13:59:51 crc kubenswrapper[4710]: I1002 13:59:51.548626 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:51 crc kubenswrapper[4710]: I1002 13:59:51.549210 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q646l" Oct 02 13:59:51 crc kubenswrapper[4710]: I1002 13:59:51.606965 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q646l" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.170916 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq"] Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.173701 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.177553 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.177739 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.180912 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq"] Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.353589 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68f92\" (UniqueName: \"kubernetes.io/projected/e8669de3-ace4-4ed7-a39f-517134c6aeaa-kube-api-access-68f92\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.354206 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8669de3-ace4-4ed7-a39f-517134c6aeaa-config-volume\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.354362 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8669de3-ace4-4ed7-a39f-517134c6aeaa-secret-volume\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.456285 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8669de3-ace4-4ed7-a39f-517134c6aeaa-secret-volume\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.456364 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68f92\" (UniqueName: \"kubernetes.io/projected/e8669de3-ace4-4ed7-a39f-517134c6aeaa-kube-api-access-68f92\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.456624 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8669de3-ace4-4ed7-a39f-517134c6aeaa-config-volume\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.457681 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8669de3-ace4-4ed7-a39f-517134c6aeaa-config-volume\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.468053 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8669de3-ace4-4ed7-a39f-517134c6aeaa-secret-volume\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.485128 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68f92\" (UniqueName: \"kubernetes.io/projected/e8669de3-ace4-4ed7-a39f-517134c6aeaa-kube-api-access-68f92\") pod \"collect-profiles-29323560-p6nkq\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:00 crc kubenswrapper[4710]: I1002 14:00:00.511956 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:01 crc kubenswrapper[4710]: I1002 14:00:01.003869 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq"] Oct 02 14:00:01 crc kubenswrapper[4710]: I1002 14:00:01.642942 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q646l" Oct 02 14:00:01 crc kubenswrapper[4710]: I1002 14:00:01.688157 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q646l"] Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.054429 4710 generic.go:334] "Generic (PLEG): container finished" podID="e8669de3-ace4-4ed7-a39f-517134c6aeaa" containerID="1bd84aebeef1a06c32fc69b747e09712160a4877a74179ca72ace34edeab2bf7" exitCode=0 Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.054503 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" event={"ID":"e8669de3-ace4-4ed7-a39f-517134c6aeaa","Type":"ContainerDied","Data":"1bd84aebeef1a06c32fc69b747e09712160a4877a74179ca72ace34edeab2bf7"} Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.054983 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" event={"ID":"e8669de3-ace4-4ed7-a39f-517134c6aeaa","Type":"ContainerStarted","Data":"ba976f9b65a3a6e6b89c74a7214ae5e44947a08b6cab36d85943a2ac8d03f7e9"} Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.055121 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q646l" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="registry-server" containerID="cri-o://a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59" gracePeriod=2 Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.592715 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q646l" Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.700688 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbqp5\" (UniqueName: \"kubernetes.io/projected/18853a2b-28e7-45b8-b0f8-92343a7035eb-kube-api-access-vbqp5\") pod \"18853a2b-28e7-45b8-b0f8-92343a7035eb\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.700902 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-utilities\") pod \"18853a2b-28e7-45b8-b0f8-92343a7035eb\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.700967 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-catalog-content\") pod \"18853a2b-28e7-45b8-b0f8-92343a7035eb\" (UID: \"18853a2b-28e7-45b8-b0f8-92343a7035eb\") " Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.701831 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-utilities" (OuterVolumeSpecName: "utilities") pod "18853a2b-28e7-45b8-b0f8-92343a7035eb" (UID: "18853a2b-28e7-45b8-b0f8-92343a7035eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.707316 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18853a2b-28e7-45b8-b0f8-92343a7035eb-kube-api-access-vbqp5" (OuterVolumeSpecName: "kube-api-access-vbqp5") pod "18853a2b-28e7-45b8-b0f8-92343a7035eb" (UID: "18853a2b-28e7-45b8-b0f8-92343a7035eb"). InnerVolumeSpecName "kube-api-access-vbqp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.752521 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18853a2b-28e7-45b8-b0f8-92343a7035eb" (UID: "18853a2b-28e7-45b8-b0f8-92343a7035eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.803729 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbqp5\" (UniqueName: \"kubernetes.io/projected/18853a2b-28e7-45b8-b0f8-92343a7035eb-kube-api-access-vbqp5\") on node \"crc\" DevicePath \"\"" Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.803976 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:00:02 crc kubenswrapper[4710]: I1002 14:00:02.804041 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18853a2b-28e7-45b8-b0f8-92343a7035eb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.068070 4710 generic.go:334] "Generic (PLEG): container finished" podID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerID="a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59" exitCode=0 Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.068126 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerDied","Data":"a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59"} Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.068231 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q646l" event={"ID":"18853a2b-28e7-45b8-b0f8-92343a7035eb","Type":"ContainerDied","Data":"d26f7480d4b19929b91fa15ac92b6fe850dedd7248463f3f8bb8ac3eb95dbe06"} Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.068149 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q646l" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.068283 4710 scope.go:117] "RemoveContainer" containerID="a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.094076 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q646l"] Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.096968 4710 scope.go:117] "RemoveContainer" containerID="19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.102172 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q646l"] Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.121544 4710 scope.go:117] "RemoveContainer" containerID="f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.168383 4710 scope.go:117] "RemoveContainer" containerID="a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59" Oct 02 14:00:03 crc kubenswrapper[4710]: E1002 14:00:03.168884 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59\": container with ID starting with a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59 not found: ID does not exist" containerID="a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.168912 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59"} err="failed to get container status \"a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59\": rpc error: code = NotFound desc = could not find container \"a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59\": container with ID starting with a0849434c920502435de24ce56c9f86ab764fff74346a080e978e397fd6a4f59 not found: ID does not exist" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.168932 4710 scope.go:117] "RemoveContainer" containerID="19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8" Oct 02 14:00:03 crc kubenswrapper[4710]: E1002 14:00:03.169362 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8\": container with ID starting with 19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8 not found: ID does not exist" containerID="19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.169386 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8"} err="failed to get container status \"19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8\": rpc error: code = NotFound desc = could not find container \"19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8\": container with ID starting with 19f99dc8deb65c3ca7d208a90a79898ff22fbcf41f527947ba718a88c03234c8 not found: ID does not exist" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.169399 4710 scope.go:117] "RemoveContainer" containerID="f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215" Oct 02 14:00:03 crc kubenswrapper[4710]: E1002 14:00:03.169590 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215\": container with ID starting with f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215 not found: ID does not exist" containerID="f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.169609 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215"} err="failed to get container status \"f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215\": rpc error: code = NotFound desc = could not find container \"f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215\": container with ID starting with f592b83b91259ec0bce239bf5d9e4f73592d4ce4536c2e76fddf7e51d5c24215 not found: ID does not exist" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.429811 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.516777 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68f92\" (UniqueName: \"kubernetes.io/projected/e8669de3-ace4-4ed7-a39f-517134c6aeaa-kube-api-access-68f92\") pod \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.516939 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8669de3-ace4-4ed7-a39f-517134c6aeaa-config-volume\") pod \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.517020 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8669de3-ace4-4ed7-a39f-517134c6aeaa-secret-volume\") pod \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\" (UID: \"e8669de3-ace4-4ed7-a39f-517134c6aeaa\") " Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.517882 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8669de3-ace4-4ed7-a39f-517134c6aeaa-config-volume" (OuterVolumeSpecName: "config-volume") pod "e8669de3-ace4-4ed7-a39f-517134c6aeaa" (UID: "e8669de3-ace4-4ed7-a39f-517134c6aeaa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.523422 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8669de3-ace4-4ed7-a39f-517134c6aeaa-kube-api-access-68f92" (OuterVolumeSpecName: "kube-api-access-68f92") pod "e8669de3-ace4-4ed7-a39f-517134c6aeaa" (UID: "e8669de3-ace4-4ed7-a39f-517134c6aeaa"). InnerVolumeSpecName "kube-api-access-68f92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.524177 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8669de3-ace4-4ed7-a39f-517134c6aeaa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e8669de3-ace4-4ed7-a39f-517134c6aeaa" (UID: "e8669de3-ace4-4ed7-a39f-517134c6aeaa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.638509 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8669de3-ace4-4ed7-a39f-517134c6aeaa-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.638593 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8669de3-ace4-4ed7-a39f-517134c6aeaa-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 14:00:03 crc kubenswrapper[4710]: I1002 14:00:03.638620 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68f92\" (UniqueName: \"kubernetes.io/projected/e8669de3-ace4-4ed7-a39f-517134c6aeaa-kube-api-access-68f92\") on node \"crc\" DevicePath \"\"" Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.086480 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" event={"ID":"e8669de3-ace4-4ed7-a39f-517134c6aeaa","Type":"ContainerDied","Data":"ba976f9b65a3a6e6b89c74a7214ae5e44947a08b6cab36d85943a2ac8d03f7e9"} Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.086552 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba976f9b65a3a6e6b89c74a7214ae5e44947a08b6cab36d85943a2ac8d03f7e9" Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.087659 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323560-p6nkq" Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.516917 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56"] Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.526651 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323515-n2t56"] Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.923681 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" path="/var/lib/kubelet/pods/18853a2b-28e7-45b8-b0f8-92343a7035eb/volumes" Oct 02 14:00:04 crc kubenswrapper[4710]: I1002 14:00:04.924449 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5569de9b-ca9d-4d55-a762-268ba7881c0b" path="/var/lib/kubelet/pods/5569de9b-ca9d-4d55-a762-268ba7881c0b/volumes" Oct 02 14:00:48 crc kubenswrapper[4710]: I1002 14:00:48.414651 4710 scope.go:117] "RemoveContainer" containerID="162c8a09c0856d03a3499f0d3854d21da87d4924247b41cb13568aba073aed58" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.149500 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323561-glh7z"] Oct 02 14:01:00 crc kubenswrapper[4710]: E1002 14:01:00.150616 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="registry-server" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.150630 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="registry-server" Oct 02 14:01:00 crc kubenswrapper[4710]: E1002 14:01:00.150693 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="extract-content" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.150700 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="extract-content" Oct 02 14:01:00 crc kubenswrapper[4710]: E1002 14:01:00.150759 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8669de3-ace4-4ed7-a39f-517134c6aeaa" containerName="collect-profiles" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.150768 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8669de3-ace4-4ed7-a39f-517134c6aeaa" containerName="collect-profiles" Oct 02 14:01:00 crc kubenswrapper[4710]: E1002 14:01:00.150782 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="extract-utilities" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.150789 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="extract-utilities" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.150992 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="18853a2b-28e7-45b8-b0f8-92343a7035eb" containerName="registry-server" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.151020 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8669de3-ace4-4ed7-a39f-517134c6aeaa" containerName="collect-profiles" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.151611 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.186890 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323561-glh7z"] Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.311727 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-fernet-keys\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.312150 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-config-data\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.312194 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnr84\" (UniqueName: \"kubernetes.io/projected/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-kube-api-access-hnr84\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.312528 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-combined-ca-bundle\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.413663 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-combined-ca-bundle\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.413766 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-fernet-keys\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.413842 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-config-data\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.413881 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnr84\" (UniqueName: \"kubernetes.io/projected/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-kube-api-access-hnr84\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.422134 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-fernet-keys\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.422333 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-combined-ca-bundle\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.424486 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-config-data\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.431827 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnr84\" (UniqueName: \"kubernetes.io/projected/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-kube-api-access-hnr84\") pod \"keystone-cron-29323561-glh7z\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:00 crc kubenswrapper[4710]: I1002 14:01:00.487628 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:01 crc kubenswrapper[4710]: I1002 14:01:01.099140 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323561-glh7z"] Oct 02 14:01:01 crc kubenswrapper[4710]: W1002 14:01:01.103869 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e87cc78_a8e9_4ba8_bfbd_f956aa4d0e56.slice/crio-1bc95da0558077df92b39e8f1135317381de08f6fe15007351b0b6c259b1fe6b WatchSource:0}: Error finding container 1bc95da0558077df92b39e8f1135317381de08f6fe15007351b0b6c259b1fe6b: Status 404 returned error can't find the container with id 1bc95da0558077df92b39e8f1135317381de08f6fe15007351b0b6c259b1fe6b Oct 02 14:01:01 crc kubenswrapper[4710]: I1002 14:01:01.746896 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323561-glh7z" event={"ID":"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56","Type":"ContainerStarted","Data":"433c4babae2e72c8b12bc52be5fdf8f595ca2a4f1d36a2d3841c27de226c888e"} Oct 02 14:01:01 crc kubenswrapper[4710]: I1002 14:01:01.747434 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323561-glh7z" event={"ID":"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56","Type":"ContainerStarted","Data":"1bc95da0558077df92b39e8f1135317381de08f6fe15007351b0b6c259b1fe6b"} Oct 02 14:01:01 crc kubenswrapper[4710]: I1002 14:01:01.779131 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323561-glh7z" podStartSLOduration=1.779106946 podStartE2EDuration="1.779106946s" podCreationTimestamp="2025-10-02 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 14:01:01.772561991 +0000 UTC m=+4305.878972874" watchObservedRunningTime="2025-10-02 14:01:01.779106946 +0000 UTC m=+4305.885517839" Oct 02 14:01:05 crc kubenswrapper[4710]: I1002 14:01:05.785530 4710 generic.go:334] "Generic (PLEG): container finished" podID="4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" containerID="433c4babae2e72c8b12bc52be5fdf8f595ca2a4f1d36a2d3841c27de226c888e" exitCode=0 Oct 02 14:01:05 crc kubenswrapper[4710]: I1002 14:01:05.785613 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323561-glh7z" event={"ID":"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56","Type":"ContainerDied","Data":"433c4babae2e72c8b12bc52be5fdf8f595ca2a4f1d36a2d3841c27de226c888e"} Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.378953 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.562880 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnr84\" (UniqueName: \"kubernetes.io/projected/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-kube-api-access-hnr84\") pod \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.563133 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-fernet-keys\") pod \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.563177 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-combined-ca-bundle\") pod \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.563227 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-config-data\") pod \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\" (UID: \"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56\") " Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.573180 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" (UID: "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.574109 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-kube-api-access-hnr84" (OuterVolumeSpecName: "kube-api-access-hnr84") pod "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" (UID: "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56"). InnerVolumeSpecName "kube-api-access-hnr84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.622146 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" (UID: "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.665795 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnr84\" (UniqueName: \"kubernetes.io/projected/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-kube-api-access-hnr84\") on node \"crc\" DevicePath \"\"" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.665851 4710 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.665872 4710 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.672761 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-config-data" (OuterVolumeSpecName: "config-data") pod "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" (UID: "4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.768059 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.808005 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323561-glh7z" event={"ID":"4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56","Type":"ContainerDied","Data":"1bc95da0558077df92b39e8f1135317381de08f6fe15007351b0b6c259b1fe6b"} Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.808230 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bc95da0558077df92b39e8f1135317381de08f6fe15007351b0b6c259b1fe6b" Oct 02 14:01:07 crc kubenswrapper[4710]: I1002 14:01:07.808052 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323561-glh7z" Oct 02 14:01:22 crc kubenswrapper[4710]: I1002 14:01:22.529861 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:01:22 crc kubenswrapper[4710]: I1002 14:01:22.530378 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:01:52 crc kubenswrapper[4710]: I1002 14:01:52.530386 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:01:52 crc kubenswrapper[4710]: I1002 14:01:52.532149 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:02:22 crc kubenswrapper[4710]: I1002 14:02:22.530468 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:02:22 crc kubenswrapper[4710]: I1002 14:02:22.530962 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:02:22 crc kubenswrapper[4710]: I1002 14:02:22.531010 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 14:02:22 crc kubenswrapper[4710]: I1002 14:02:22.531639 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 14:02:22 crc kubenswrapper[4710]: I1002 14:02:22.531690 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" gracePeriod=600 Oct 02 14:02:23 crc kubenswrapper[4710]: E1002 14:02:23.566543 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:02:23 crc kubenswrapper[4710]: I1002 14:02:23.754771 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" exitCode=0 Oct 02 14:02:23 crc kubenswrapper[4710]: I1002 14:02:23.754828 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b"} Oct 02 14:02:23 crc kubenswrapper[4710]: I1002 14:02:23.754875 4710 scope.go:117] "RemoveContainer" containerID="357899fea19c2894c442c75e131da52c0455c786dbcf957b2a9376ad405bb029" Oct 02 14:02:23 crc kubenswrapper[4710]: I1002 14:02:23.755786 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:02:23 crc kubenswrapper[4710]: E1002 14:02:23.756136 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:02:36 crc kubenswrapper[4710]: I1002 14:02:36.918245 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:02:36 crc kubenswrapper[4710]: E1002 14:02:36.919208 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:02:47 crc kubenswrapper[4710]: I1002 14:02:47.904826 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:02:47 crc kubenswrapper[4710]: E1002 14:02:47.905625 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:02:58 crc kubenswrapper[4710]: I1002 14:02:58.905189 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:02:58 crc kubenswrapper[4710]: E1002 14:02:58.906240 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:03:10 crc kubenswrapper[4710]: I1002 14:03:10.904838 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:03:10 crc kubenswrapper[4710]: E1002 14:03:10.906440 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:03:22 crc kubenswrapper[4710]: I1002 14:03:22.904529 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:03:22 crc kubenswrapper[4710]: E1002 14:03:22.905484 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:03:36 crc kubenswrapper[4710]: I1002 14:03:36.916032 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:03:36 crc kubenswrapper[4710]: E1002 14:03:36.916905 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:03:48 crc kubenswrapper[4710]: I1002 14:03:48.909089 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:03:48 crc kubenswrapper[4710]: E1002 14:03:48.910166 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:03:59 crc kubenswrapper[4710]: I1002 14:03:59.905452 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:03:59 crc kubenswrapper[4710]: E1002 14:03:59.906414 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:04:11 crc kubenswrapper[4710]: I1002 14:04:11.904475 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:04:11 crc kubenswrapper[4710]: E1002 14:04:11.905261 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:04:22 crc kubenswrapper[4710]: I1002 14:04:22.905915 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:04:22 crc kubenswrapper[4710]: E1002 14:04:22.907135 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:04:35 crc kubenswrapper[4710]: I1002 14:04:35.906395 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:04:35 crc kubenswrapper[4710]: E1002 14:04:35.907795 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:04:47 crc kubenswrapper[4710]: I1002 14:04:47.907697 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:04:47 crc kubenswrapper[4710]: E1002 14:04:47.910788 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:04:58 crc kubenswrapper[4710]: I1002 14:04:58.904821 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:04:58 crc kubenswrapper[4710]: E1002 14:04:58.905652 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:05:11 crc kubenswrapper[4710]: I1002 14:05:11.904159 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:05:11 crc kubenswrapper[4710]: E1002 14:05:11.904954 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:05:26 crc kubenswrapper[4710]: I1002 14:05:26.904513 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:05:26 crc kubenswrapper[4710]: E1002 14:05:26.905216 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:05:37 crc kubenswrapper[4710]: I1002 14:05:37.905589 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:05:37 crc kubenswrapper[4710]: E1002 14:05:37.906782 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:05:50 crc kubenswrapper[4710]: I1002 14:05:50.905215 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:05:50 crc kubenswrapper[4710]: E1002 14:05:50.906120 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:06:01 crc kubenswrapper[4710]: I1002 14:06:01.904800 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:06:01 crc kubenswrapper[4710]: E1002 14:06:01.905570 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:06:15 crc kubenswrapper[4710]: I1002 14:06:15.906100 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:06:15 crc kubenswrapper[4710]: E1002 14:06:15.907721 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:06:15 crc kubenswrapper[4710]: I1002 14:06:15.994228 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5tcc"] Oct 02 14:06:15 crc kubenswrapper[4710]: E1002 14:06:15.998576 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" containerName="keystone-cron" Oct 02 14:06:15 crc kubenswrapper[4710]: I1002 14:06:15.998620 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" containerName="keystone-cron" Oct 02 14:06:15 crc kubenswrapper[4710]: I1002 14:06:15.999036 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56" containerName="keystone-cron" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.001360 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.013604 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5tcc"] Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.125047 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwhqh\" (UniqueName: \"kubernetes.io/projected/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-kube-api-access-qwhqh\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.125211 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-utilities\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.125254 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-catalog-content\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.227422 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-utilities\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.227477 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-catalog-content\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.227601 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwhqh\" (UniqueName: \"kubernetes.io/projected/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-kube-api-access-qwhqh\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.228087 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-utilities\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.228145 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-catalog-content\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.244896 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwhqh\" (UniqueName: \"kubernetes.io/projected/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-kube-api-access-qwhqh\") pod \"redhat-operators-g5tcc\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.322353 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:16 crc kubenswrapper[4710]: I1002 14:06:16.776521 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5tcc"] Oct 02 14:06:17 crc kubenswrapper[4710]: I1002 14:06:17.412441 4710 generic.go:334] "Generic (PLEG): container finished" podID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerID="a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26" exitCode=0 Oct 02 14:06:17 crc kubenswrapper[4710]: I1002 14:06:17.412615 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerDied","Data":"a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26"} Oct 02 14:06:17 crc kubenswrapper[4710]: I1002 14:06:17.412811 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerStarted","Data":"84544df5f44e9f93235bf2965f005921a2c9b3694bd696fd8b3f11f63320fb56"} Oct 02 14:06:17 crc kubenswrapper[4710]: I1002 14:06:17.415713 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 14:06:19 crc kubenswrapper[4710]: I1002 14:06:19.437728 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerStarted","Data":"92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076"} Oct 02 14:06:22 crc kubenswrapper[4710]: I1002 14:06:22.479383 4710 generic.go:334] "Generic (PLEG): container finished" podID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerID="92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076" exitCode=0 Oct 02 14:06:22 crc kubenswrapper[4710]: I1002 14:06:22.479544 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerDied","Data":"92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076"} Oct 02 14:06:23 crc kubenswrapper[4710]: I1002 14:06:23.491721 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerStarted","Data":"0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7"} Oct 02 14:06:23 crc kubenswrapper[4710]: I1002 14:06:23.524933 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5tcc" podStartSLOduration=3.043902844 podStartE2EDuration="8.524903519s" podCreationTimestamp="2025-10-02 14:06:15 +0000 UTC" firstStartedPulling="2025-10-02 14:06:17.415326849 +0000 UTC m=+4621.521737752" lastFinishedPulling="2025-10-02 14:06:22.896327514 +0000 UTC m=+4627.002738427" observedRunningTime="2025-10-02 14:06:23.51265289 +0000 UTC m=+4627.619063793" watchObservedRunningTime="2025-10-02 14:06:23.524903519 +0000 UTC m=+4627.631314432" Oct 02 14:06:26 crc kubenswrapper[4710]: I1002 14:06:26.322813 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:26 crc kubenswrapper[4710]: I1002 14:06:26.323470 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:27 crc kubenswrapper[4710]: I1002 14:06:27.410913 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g5tcc" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="registry-server" probeResult="failure" output=< Oct 02 14:06:27 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 14:06:27 crc kubenswrapper[4710]: > Oct 02 14:06:30 crc kubenswrapper[4710]: I1002 14:06:30.905054 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:06:30 crc kubenswrapper[4710]: E1002 14:06:30.905768 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:06:36 crc kubenswrapper[4710]: I1002 14:06:36.400413 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:36 crc kubenswrapper[4710]: I1002 14:06:36.479573 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:36 crc kubenswrapper[4710]: I1002 14:06:36.652586 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5tcc"] Oct 02 14:06:37 crc kubenswrapper[4710]: I1002 14:06:37.667666 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g5tcc" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="registry-server" containerID="cri-o://0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7" gracePeriod=2 Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.195454 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.371557 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwhqh\" (UniqueName: \"kubernetes.io/projected/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-kube-api-access-qwhqh\") pod \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.372252 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-catalog-content\") pod \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.372296 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-utilities\") pod \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\" (UID: \"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b\") " Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.373451 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-utilities" (OuterVolumeSpecName: "utilities") pod "dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" (UID: "dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.379702 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-kube-api-access-qwhqh" (OuterVolumeSpecName: "kube-api-access-qwhqh") pod "dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" (UID: "dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b"). InnerVolumeSpecName "kube-api-access-qwhqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.474675 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.474705 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwhqh\" (UniqueName: \"kubernetes.io/projected/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-kube-api-access-qwhqh\") on node \"crc\" DevicePath \"\"" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.512917 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" (UID: "dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.576120 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.682990 4710 generic.go:334] "Generic (PLEG): container finished" podID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerID="0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7" exitCode=0 Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.683041 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerDied","Data":"0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7"} Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.683070 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5tcc" event={"ID":"dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b","Type":"ContainerDied","Data":"84544df5f44e9f93235bf2965f005921a2c9b3694bd696fd8b3f11f63320fb56"} Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.683081 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5tcc" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.683088 4710 scope.go:117] "RemoveContainer" containerID="0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.719892 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5tcc"] Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.730537 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g5tcc"] Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.743795 4710 scope.go:117] "RemoveContainer" containerID="92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.777890 4710 scope.go:117] "RemoveContainer" containerID="a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.824711 4710 scope.go:117] "RemoveContainer" containerID="0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7" Oct 02 14:06:38 crc kubenswrapper[4710]: E1002 14:06:38.825171 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7\": container with ID starting with 0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7 not found: ID does not exist" containerID="0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.825226 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7"} err="failed to get container status \"0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7\": rpc error: code = NotFound desc = could not find container \"0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7\": container with ID starting with 0f39b5c752c62f353de8b0da8b506bc13b30bd16e91018759efc431e9cd108b7 not found: ID does not exist" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.825260 4710 scope.go:117] "RemoveContainer" containerID="92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076" Oct 02 14:06:38 crc kubenswrapper[4710]: E1002 14:06:38.825616 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076\": container with ID starting with 92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076 not found: ID does not exist" containerID="92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.825649 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076"} err="failed to get container status \"92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076\": rpc error: code = NotFound desc = could not find container \"92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076\": container with ID starting with 92db2275ea5dc687577d739921b283844a68a451997e4dd1f247e85cf7afa076 not found: ID does not exist" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.825669 4710 scope.go:117] "RemoveContainer" containerID="a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26" Oct 02 14:06:38 crc kubenswrapper[4710]: E1002 14:06:38.825990 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26\": container with ID starting with a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26 not found: ID does not exist" containerID="a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.826013 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26"} err="failed to get container status \"a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26\": rpc error: code = NotFound desc = could not find container \"a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26\": container with ID starting with a6dec0b65c014e516823e735928786bafb3b689d53cd3ad5e0b1735648d45c26 not found: ID does not exist" Oct 02 14:06:38 crc kubenswrapper[4710]: I1002 14:06:38.916631 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" path="/var/lib/kubelet/pods/dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b/volumes" Oct 02 14:06:41 crc kubenswrapper[4710]: I1002 14:06:41.904941 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:06:41 crc kubenswrapper[4710]: E1002 14:06:41.905507 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:06:52 crc kubenswrapper[4710]: I1002 14:06:52.905128 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:06:52 crc kubenswrapper[4710]: E1002 14:06:52.905809 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:07:07 crc kubenswrapper[4710]: I1002 14:07:07.905190 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:07:07 crc kubenswrapper[4710]: E1002 14:07:07.906098 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:07:22 crc kubenswrapper[4710]: I1002 14:07:22.905105 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:07:22 crc kubenswrapper[4710]: E1002 14:07:22.905835 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:07:34 crc kubenswrapper[4710]: I1002 14:07:34.904845 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:07:36 crc kubenswrapper[4710]: I1002 14:07:36.353937 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"5dac050993e686083e1b20362dbad1ed830cf986f45b609b62f828e97962bb8a"} Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.009803 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4clgv"] Oct 02 14:09:03 crc kubenswrapper[4710]: E1002 14:09:03.011012 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="extract-content" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.011036 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="extract-content" Oct 02 14:09:03 crc kubenswrapper[4710]: E1002 14:09:03.011063 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="registry-server" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.011076 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="registry-server" Oct 02 14:09:03 crc kubenswrapper[4710]: E1002 14:09:03.011125 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="extract-utilities" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.011138 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="extract-utilities" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.011579 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd5e8aab-237f-4b0b-bbe2-2a6c098ed64b" containerName="registry-server" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.014173 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.023083 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4clgv"] Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.105715 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-utilities\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.105793 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-catalog-content\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.105926 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v4fh\" (UniqueName: \"kubernetes.io/projected/b66094b6-5d54-4074-97bb-311538b04c91-kube-api-access-8v4fh\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.208172 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v4fh\" (UniqueName: \"kubernetes.io/projected/b66094b6-5d54-4074-97bb-311538b04c91-kube-api-access-8v4fh\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.208331 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-utilities\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.208365 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-catalog-content\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.209050 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-catalog-content\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.209050 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-utilities\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.230452 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v4fh\" (UniqueName: \"kubernetes.io/projected/b66094b6-5d54-4074-97bb-311538b04c91-kube-api-access-8v4fh\") pod \"certified-operators-4clgv\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.350858 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:03 crc kubenswrapper[4710]: I1002 14:09:03.855296 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4clgv"] Oct 02 14:09:04 crc kubenswrapper[4710]: I1002 14:09:04.321203 4710 generic.go:334] "Generic (PLEG): container finished" podID="b66094b6-5d54-4074-97bb-311538b04c91" containerID="c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd" exitCode=0 Oct 02 14:09:04 crc kubenswrapper[4710]: I1002 14:09:04.321278 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerDied","Data":"c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd"} Oct 02 14:09:04 crc kubenswrapper[4710]: I1002 14:09:04.321512 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerStarted","Data":"cbf3e0a33f082f8bef0f2d3b7ddcad9ee3152e3de5c056568c1dbf2d44ba57df"} Oct 02 14:09:06 crc kubenswrapper[4710]: I1002 14:09:06.347614 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerStarted","Data":"d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628"} Oct 02 14:09:08 crc kubenswrapper[4710]: I1002 14:09:08.372350 4710 generic.go:334] "Generic (PLEG): container finished" podID="b66094b6-5d54-4074-97bb-311538b04c91" containerID="d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628" exitCode=0 Oct 02 14:09:08 crc kubenswrapper[4710]: I1002 14:09:08.372431 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerDied","Data":"d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628"} Oct 02 14:09:10 crc kubenswrapper[4710]: I1002 14:09:10.396307 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerStarted","Data":"8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9"} Oct 02 14:09:10 crc kubenswrapper[4710]: I1002 14:09:10.424871 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4clgv" podStartSLOduration=3.577437867 podStartE2EDuration="8.424853773s" podCreationTimestamp="2025-10-02 14:09:02 +0000 UTC" firstStartedPulling="2025-10-02 14:09:04.32356828 +0000 UTC m=+4788.429979163" lastFinishedPulling="2025-10-02 14:09:09.170984186 +0000 UTC m=+4793.277395069" observedRunningTime="2025-10-02 14:09:10.419618691 +0000 UTC m=+4794.526029584" watchObservedRunningTime="2025-10-02 14:09:10.424853773 +0000 UTC m=+4794.531264656" Oct 02 14:09:13 crc kubenswrapper[4710]: I1002 14:09:13.351288 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:13 crc kubenswrapper[4710]: I1002 14:09:13.352791 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:13 crc kubenswrapper[4710]: I1002 14:09:13.414272 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:14 crc kubenswrapper[4710]: I1002 14:09:14.504623 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:14 crc kubenswrapper[4710]: I1002 14:09:14.581421 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4clgv"] Oct 02 14:09:16 crc kubenswrapper[4710]: I1002 14:09:16.461434 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4clgv" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="registry-server" containerID="cri-o://8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9" gracePeriod=2 Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.058538 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.068169 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-catalog-content\") pod \"b66094b6-5d54-4074-97bb-311538b04c91\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.126866 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b66094b6-5d54-4074-97bb-311538b04c91" (UID: "b66094b6-5d54-4074-97bb-311538b04c91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.169558 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-utilities\") pod \"b66094b6-5d54-4074-97bb-311538b04c91\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.170324 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-utilities" (OuterVolumeSpecName: "utilities") pod "b66094b6-5d54-4074-97bb-311538b04c91" (UID: "b66094b6-5d54-4074-97bb-311538b04c91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.170342 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v4fh\" (UniqueName: \"kubernetes.io/projected/b66094b6-5d54-4074-97bb-311538b04c91-kube-api-access-8v4fh\") pod \"b66094b6-5d54-4074-97bb-311538b04c91\" (UID: \"b66094b6-5d54-4074-97bb-311538b04c91\") " Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.171660 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.172102 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66094b6-5d54-4074-97bb-311538b04c91-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.176103 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66094b6-5d54-4074-97bb-311538b04c91-kube-api-access-8v4fh" (OuterVolumeSpecName: "kube-api-access-8v4fh") pod "b66094b6-5d54-4074-97bb-311538b04c91" (UID: "b66094b6-5d54-4074-97bb-311538b04c91"). InnerVolumeSpecName "kube-api-access-8v4fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.274390 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v4fh\" (UniqueName: \"kubernetes.io/projected/b66094b6-5d54-4074-97bb-311538b04c91-kube-api-access-8v4fh\") on node \"crc\" DevicePath \"\"" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.477737 4710 generic.go:334] "Generic (PLEG): container finished" podID="b66094b6-5d54-4074-97bb-311538b04c91" containerID="8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9" exitCode=0 Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.477827 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerDied","Data":"8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9"} Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.478203 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4clgv" event={"ID":"b66094b6-5d54-4074-97bb-311538b04c91","Type":"ContainerDied","Data":"cbf3e0a33f082f8bef0f2d3b7ddcad9ee3152e3de5c056568c1dbf2d44ba57df"} Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.478237 4710 scope.go:117] "RemoveContainer" containerID="8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.477972 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4clgv" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.516892 4710 scope.go:117] "RemoveContainer" containerID="d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.546705 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4clgv"] Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.565421 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4clgv"] Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.573316 4710 scope.go:117] "RemoveContainer" containerID="c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.616854 4710 scope.go:117] "RemoveContainer" containerID="8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9" Oct 02 14:09:17 crc kubenswrapper[4710]: E1002 14:09:17.617441 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9\": container with ID starting with 8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9 not found: ID does not exist" containerID="8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.617637 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9"} err="failed to get container status \"8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9\": rpc error: code = NotFound desc = could not find container \"8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9\": container with ID starting with 8a215529059c1989a5efa5665d5dd624f3f7e66f8636cf3f6d0c6e506ba0b0f9 not found: ID does not exist" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.617903 4710 scope.go:117] "RemoveContainer" containerID="d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628" Oct 02 14:09:17 crc kubenswrapper[4710]: E1002 14:09:17.620330 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628\": container with ID starting with d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628 not found: ID does not exist" containerID="d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.620434 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628"} err="failed to get container status \"d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628\": rpc error: code = NotFound desc = could not find container \"d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628\": container with ID starting with d3e91d85c23351ef68ffb669cb5623624d20bf1e89f9521b915fcec41791f628 not found: ID does not exist" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.620518 4710 scope.go:117] "RemoveContainer" containerID="c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd" Oct 02 14:09:17 crc kubenswrapper[4710]: E1002 14:09:17.621019 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd\": container with ID starting with c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd not found: ID does not exist" containerID="c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd" Oct 02 14:09:17 crc kubenswrapper[4710]: I1002 14:09:17.621189 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd"} err="failed to get container status \"c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd\": rpc error: code = NotFound desc = could not find container \"c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd\": container with ID starting with c78491e753fc55b89a8291fe0dbfb50dc5dc0ba51b67a8f43c74fcf70599affd not found: ID does not exist" Oct 02 14:09:18 crc kubenswrapper[4710]: I1002 14:09:18.919159 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b66094b6-5d54-4074-97bb-311538b04c91" path="/var/lib/kubelet/pods/b66094b6-5d54-4074-97bb-311538b04c91/volumes" Oct 02 14:09:52 crc kubenswrapper[4710]: I1002 14:09:52.530308 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:09:52 crc kubenswrapper[4710]: I1002 14:09:52.532218 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:10:22 crc kubenswrapper[4710]: I1002 14:10:22.530482 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:10:22 crc kubenswrapper[4710]: I1002 14:10:22.531147 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:10:52 crc kubenswrapper[4710]: I1002 14:10:52.530158 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:10:52 crc kubenswrapper[4710]: I1002 14:10:52.530733 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:10:52 crc kubenswrapper[4710]: I1002 14:10:52.530812 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 14:10:52 crc kubenswrapper[4710]: I1002 14:10:52.531505 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5dac050993e686083e1b20362dbad1ed830cf986f45b609b62f828e97962bb8a"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 14:10:52 crc kubenswrapper[4710]: I1002 14:10:52.531576 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://5dac050993e686083e1b20362dbad1ed830cf986f45b609b62f828e97962bb8a" gracePeriod=600 Oct 02 14:10:53 crc kubenswrapper[4710]: I1002 14:10:53.655657 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="5dac050993e686083e1b20362dbad1ed830cf986f45b609b62f828e97962bb8a" exitCode=0 Oct 02 14:10:53 crc kubenswrapper[4710]: I1002 14:10:53.655709 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"5dac050993e686083e1b20362dbad1ed830cf986f45b609b62f828e97962bb8a"} Oct 02 14:10:53 crc kubenswrapper[4710]: I1002 14:10:53.656902 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10"} Oct 02 14:10:53 crc kubenswrapper[4710]: I1002 14:10:53.656959 4710 scope.go:117] "RemoveContainer" containerID="8b11fafcd92051453c77c2bfc6e41686448010e04e023bec0354c792ebc2401b" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.600715 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bzlvx"] Oct 02 14:11:08 crc kubenswrapper[4710]: E1002 14:11:08.602463 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="extract-content" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.602496 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="extract-content" Oct 02 14:11:08 crc kubenswrapper[4710]: E1002 14:11:08.602553 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="extract-utilities" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.602572 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="extract-utilities" Oct 02 14:11:08 crc kubenswrapper[4710]: E1002 14:11:08.602636 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="registry-server" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.602655 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="registry-server" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.603220 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66094b6-5d54-4074-97bb-311538b04c91" containerName="registry-server" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.606887 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.620876 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bzlvx"] Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.773924 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl7cg\" (UniqueName: \"kubernetes.io/projected/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-kube-api-access-kl7cg\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.774542 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-catalog-content\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.774725 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-utilities\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.877029 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl7cg\" (UniqueName: \"kubernetes.io/projected/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-kube-api-access-kl7cg\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.877106 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-catalog-content\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.877197 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-utilities\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.877854 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-catalog-content\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.880920 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-utilities\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.915105 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl7cg\" (UniqueName: \"kubernetes.io/projected/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-kube-api-access-kl7cg\") pod \"community-operators-bzlvx\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:08 crc kubenswrapper[4710]: I1002 14:11:08.963282 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:09 crc kubenswrapper[4710]: W1002 14:11:09.530562 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e71c3a_a7b0_42b3_9f96_cd6245f6802a.slice/crio-e9bacdb25597150dbd426525d7d9091a2dfb5db4d2b141a0d4710052d9039b61 WatchSource:0}: Error finding container e9bacdb25597150dbd426525d7d9091a2dfb5db4d2b141a0d4710052d9039b61: Status 404 returned error can't find the container with id e9bacdb25597150dbd426525d7d9091a2dfb5db4d2b141a0d4710052d9039b61 Oct 02 14:11:09 crc kubenswrapper[4710]: I1002 14:11:09.530862 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bzlvx"] Oct 02 14:11:09 crc kubenswrapper[4710]: I1002 14:11:09.846546 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bzlvx" event={"ID":"93e71c3a-a7b0-42b3-9f96-cd6245f6802a","Type":"ContainerStarted","Data":"e9bacdb25597150dbd426525d7d9091a2dfb5db4d2b141a0d4710052d9039b61"} Oct 02 14:11:10 crc kubenswrapper[4710]: I1002 14:11:10.859818 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerID="560a26949642405e1704559aed536a6a845507ade3d2537d981df2defbb848f2" exitCode=0 Oct 02 14:11:10 crc kubenswrapper[4710]: I1002 14:11:10.860056 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bzlvx" event={"ID":"93e71c3a-a7b0-42b3-9f96-cd6245f6802a","Type":"ContainerDied","Data":"560a26949642405e1704559aed536a6a845507ade3d2537d981df2defbb848f2"} Oct 02 14:11:12 crc kubenswrapper[4710]: I1002 14:11:12.883414 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerID="12365d601c5061dff6a97ef7fbd527715222a95daca0948831f2ba881bdc341e" exitCode=0 Oct 02 14:11:12 crc kubenswrapper[4710]: I1002 14:11:12.883466 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bzlvx" event={"ID":"93e71c3a-a7b0-42b3-9f96-cd6245f6802a","Type":"ContainerDied","Data":"12365d601c5061dff6a97ef7fbd527715222a95daca0948831f2ba881bdc341e"} Oct 02 14:11:13 crc kubenswrapper[4710]: I1002 14:11:13.896007 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bzlvx" event={"ID":"93e71c3a-a7b0-42b3-9f96-cd6245f6802a","Type":"ContainerStarted","Data":"d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef"} Oct 02 14:11:13 crc kubenswrapper[4710]: I1002 14:11:13.927733 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bzlvx" podStartSLOduration=3.224626038 podStartE2EDuration="5.927703867s" podCreationTimestamp="2025-10-02 14:11:08 +0000 UTC" firstStartedPulling="2025-10-02 14:11:10.862454081 +0000 UTC m=+4914.968864964" lastFinishedPulling="2025-10-02 14:11:13.56553191 +0000 UTC m=+4917.671942793" observedRunningTime="2025-10-02 14:11:13.917127181 +0000 UTC m=+4918.023538094" watchObservedRunningTime="2025-10-02 14:11:13.927703867 +0000 UTC m=+4918.034114770" Oct 02 14:11:18 crc kubenswrapper[4710]: I1002 14:11:18.964575 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:18 crc kubenswrapper[4710]: I1002 14:11:18.965536 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:19 crc kubenswrapper[4710]: I1002 14:11:19.030391 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:21 crc kubenswrapper[4710]: I1002 14:11:21.268119 4710 trace.go:236] Trace[789737577]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-94fd2" (02-Oct-2025 14:11:20.078) (total time: 1189ms): Oct 02 14:11:21 crc kubenswrapper[4710]: Trace[789737577]: [1.189422354s] [1.189422354s] END Oct 02 14:11:21 crc kubenswrapper[4710]: I1002 14:11:21.305017 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:21 crc kubenswrapper[4710]: I1002 14:11:21.360574 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bzlvx"] Oct 02 14:11:22 crc kubenswrapper[4710]: I1002 14:11:22.990279 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bzlvx" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="registry-server" containerID="cri-o://d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef" gracePeriod=2 Oct 02 14:11:23 crc kubenswrapper[4710]: E1002 14:11:23.744318 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e71c3a_a7b0_42b3_9f96_cd6245f6802a.slice/crio-d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93e71c3a_a7b0_42b3_9f96_cd6245f6802a.slice/crio-conmon-d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef.scope\": RecentStats: unable to find data in memory cache]" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.003387 4710 generic.go:334] "Generic (PLEG): container finished" podID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerID="d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef" exitCode=0 Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.003424 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bzlvx" event={"ID":"93e71c3a-a7b0-42b3-9f96-cd6245f6802a","Type":"ContainerDied","Data":"d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef"} Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.606080 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.639531 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl7cg\" (UniqueName: \"kubernetes.io/projected/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-kube-api-access-kl7cg\") pod \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.639674 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-utilities\") pod \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.639774 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-catalog-content\") pod \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\" (UID: \"93e71c3a-a7b0-42b3-9f96-cd6245f6802a\") " Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.641597 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-utilities" (OuterVolumeSpecName: "utilities") pod "93e71c3a-a7b0-42b3-9f96-cd6245f6802a" (UID: "93e71c3a-a7b0-42b3-9f96-cd6245f6802a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.650028 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-kube-api-access-kl7cg" (OuterVolumeSpecName: "kube-api-access-kl7cg") pod "93e71c3a-a7b0-42b3-9f96-cd6245f6802a" (UID: "93e71c3a-a7b0-42b3-9f96-cd6245f6802a"). InnerVolumeSpecName "kube-api-access-kl7cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.704114 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93e71c3a-a7b0-42b3-9f96-cd6245f6802a" (UID: "93e71c3a-a7b0-42b3-9f96-cd6245f6802a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.742810 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.742854 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:11:24 crc kubenswrapper[4710]: I1002 14:11:24.742868 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl7cg\" (UniqueName: \"kubernetes.io/projected/93e71c3a-a7b0-42b3-9f96-cd6245f6802a-kube-api-access-kl7cg\") on node \"crc\" DevicePath \"\"" Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.018721 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bzlvx" event={"ID":"93e71c3a-a7b0-42b3-9f96-cd6245f6802a","Type":"ContainerDied","Data":"e9bacdb25597150dbd426525d7d9091a2dfb5db4d2b141a0d4710052d9039b61"} Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.018805 4710 scope.go:117] "RemoveContainer" containerID="d62d06d37917dc5bccf5146315f3900b834824336e6714bd56f0be8881acb4ef" Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.019847 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bzlvx" Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.050847 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bzlvx"] Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.057104 4710 scope.go:117] "RemoveContainer" containerID="12365d601c5061dff6a97ef7fbd527715222a95daca0948831f2ba881bdc341e" Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.061372 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bzlvx"] Oct 02 14:11:25 crc kubenswrapper[4710]: I1002 14:11:25.094945 4710 scope.go:117] "RemoveContainer" containerID="560a26949642405e1704559aed536a6a845507ade3d2537d981df2defbb848f2" Oct 02 14:11:26 crc kubenswrapper[4710]: I1002 14:11:26.917298 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" path="/var/lib/kubelet/pods/93e71c3a-a7b0-42b3-9f96-cd6245f6802a/volumes" Oct 02 14:12:52 crc kubenswrapper[4710]: I1002 14:12:52.530984 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:12:52 crc kubenswrapper[4710]: I1002 14:12:52.532066 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:13:22 crc kubenswrapper[4710]: I1002 14:13:22.530932 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:13:22 crc kubenswrapper[4710]: I1002 14:13:22.531681 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.530570 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.531145 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.531191 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.531949 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.531994 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" gracePeriod=600 Oct 02 14:13:52 crc kubenswrapper[4710]: E1002 14:13:52.666202 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.666629 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" exitCode=0 Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.666665 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10"} Oct 02 14:13:52 crc kubenswrapper[4710]: I1002 14:13:52.666705 4710 scope.go:117] "RemoveContainer" containerID="5dac050993e686083e1b20362dbad1ed830cf986f45b609b62f828e97962bb8a" Oct 02 14:13:53 crc kubenswrapper[4710]: I1002 14:13:53.685077 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:13:53 crc kubenswrapper[4710]: E1002 14:13:53.685885 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:14:08 crc kubenswrapper[4710]: I1002 14:14:08.904983 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:14:08 crc kubenswrapper[4710]: E1002 14:14:08.905966 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.776132 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7cc"] Oct 02 14:14:09 crc kubenswrapper[4710]: E1002 14:14:09.779367 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="extract-utilities" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.779392 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="extract-utilities" Oct 02 14:14:09 crc kubenswrapper[4710]: E1002 14:14:09.779431 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="registry-server" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.779438 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="registry-server" Oct 02 14:14:09 crc kubenswrapper[4710]: E1002 14:14:09.779468 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="extract-content" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.779475 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="extract-content" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.780993 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e71c3a-a7b0-42b3-9f96-cd6245f6802a" containerName="registry-server" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.788097 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.806759 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7cc"] Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.913982 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-catalog-content\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.914042 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-utilities\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:09 crc kubenswrapper[4710]: I1002 14:14:09.914106 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt6ws\" (UniqueName: \"kubernetes.io/projected/c4175a50-26b9-459f-9f46-840fdeeb1845-kube-api-access-tt6ws\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.015407 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-catalog-content\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.015468 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-utilities\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.015527 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt6ws\" (UniqueName: \"kubernetes.io/projected/c4175a50-26b9-459f-9f46-840fdeeb1845-kube-api-access-tt6ws\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.016013 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-catalog-content\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.016039 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-utilities\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.040726 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt6ws\" (UniqueName: \"kubernetes.io/projected/c4175a50-26b9-459f-9f46-840fdeeb1845-kube-api-access-tt6ws\") pod \"redhat-marketplace-2j7cc\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.130071 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.662233 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7cc"] Oct 02 14:14:10 crc kubenswrapper[4710]: W1002 14:14:10.672437 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4175a50_26b9_459f_9f46_840fdeeb1845.slice/crio-7d87dc1d5fe6c6c72910a5d374eea345c9dd5f5ba000b170c6667c5aefa6a91d WatchSource:0}: Error finding container 7d87dc1d5fe6c6c72910a5d374eea345c9dd5f5ba000b170c6667c5aefa6a91d: Status 404 returned error can't find the container with id 7d87dc1d5fe6c6c72910a5d374eea345c9dd5f5ba000b170c6667c5aefa6a91d Oct 02 14:14:10 crc kubenswrapper[4710]: I1002 14:14:10.864351 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7cc" event={"ID":"c4175a50-26b9-459f-9f46-840fdeeb1845","Type":"ContainerStarted","Data":"7d87dc1d5fe6c6c72910a5d374eea345c9dd5f5ba000b170c6667c5aefa6a91d"} Oct 02 14:14:11 crc kubenswrapper[4710]: I1002 14:14:11.875479 4710 generic.go:334] "Generic (PLEG): container finished" podID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerID="6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8" exitCode=0 Oct 02 14:14:11 crc kubenswrapper[4710]: I1002 14:14:11.875558 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7cc" event={"ID":"c4175a50-26b9-459f-9f46-840fdeeb1845","Type":"ContainerDied","Data":"6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8"} Oct 02 14:14:11 crc kubenswrapper[4710]: I1002 14:14:11.877755 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 14:14:13 crc kubenswrapper[4710]: I1002 14:14:13.898239 4710 generic.go:334] "Generic (PLEG): container finished" podID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerID="0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6" exitCode=0 Oct 02 14:14:13 crc kubenswrapper[4710]: I1002 14:14:13.898306 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7cc" event={"ID":"c4175a50-26b9-459f-9f46-840fdeeb1845","Type":"ContainerDied","Data":"0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6"} Oct 02 14:14:14 crc kubenswrapper[4710]: I1002 14:14:14.914921 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7cc" event={"ID":"c4175a50-26b9-459f-9f46-840fdeeb1845","Type":"ContainerStarted","Data":"25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7"} Oct 02 14:14:14 crc kubenswrapper[4710]: I1002 14:14:14.934039 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2j7cc" podStartSLOduration=3.412466985 podStartE2EDuration="5.93402328s" podCreationTimestamp="2025-10-02 14:14:09 +0000 UTC" firstStartedPulling="2025-10-02 14:14:11.877528544 +0000 UTC m=+5095.983939427" lastFinishedPulling="2025-10-02 14:14:14.399084839 +0000 UTC m=+5098.505495722" observedRunningTime="2025-10-02 14:14:14.926835809 +0000 UTC m=+5099.033246732" watchObservedRunningTime="2025-10-02 14:14:14.93402328 +0000 UTC m=+5099.040434163" Oct 02 14:14:19 crc kubenswrapper[4710]: I1002 14:14:19.905132 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:14:19 crc kubenswrapper[4710]: E1002 14:14:19.906169 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:14:20 crc kubenswrapper[4710]: I1002 14:14:20.130236 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:20 crc kubenswrapper[4710]: I1002 14:14:20.130354 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:20 crc kubenswrapper[4710]: I1002 14:14:20.177773 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:21 crc kubenswrapper[4710]: I1002 14:14:21.051401 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:21 crc kubenswrapper[4710]: I1002 14:14:21.114593 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7cc"] Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.006025 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2j7cc" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="registry-server" containerID="cri-o://25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7" gracePeriod=2 Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.581426 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.711537 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-utilities\") pod \"c4175a50-26b9-459f-9f46-840fdeeb1845\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.712742 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-utilities" (OuterVolumeSpecName: "utilities") pod "c4175a50-26b9-459f-9f46-840fdeeb1845" (UID: "c4175a50-26b9-459f-9f46-840fdeeb1845"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.712950 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt6ws\" (UniqueName: \"kubernetes.io/projected/c4175a50-26b9-459f-9f46-840fdeeb1845-kube-api-access-tt6ws\") pod \"c4175a50-26b9-459f-9f46-840fdeeb1845\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.714583 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-catalog-content\") pod \"c4175a50-26b9-459f-9f46-840fdeeb1845\" (UID: \"c4175a50-26b9-459f-9f46-840fdeeb1845\") " Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.721362 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4175a50-26b9-459f-9f46-840fdeeb1845-kube-api-access-tt6ws" (OuterVolumeSpecName: "kube-api-access-tt6ws") pod "c4175a50-26b9-459f-9f46-840fdeeb1845" (UID: "c4175a50-26b9-459f-9f46-840fdeeb1845"). InnerVolumeSpecName "kube-api-access-tt6ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.733336 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt6ws\" (UniqueName: \"kubernetes.io/projected/c4175a50-26b9-459f-9f46-840fdeeb1845-kube-api-access-tt6ws\") on node \"crc\" DevicePath \"\"" Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.733380 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.734182 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4175a50-26b9-459f-9f46-840fdeeb1845" (UID: "c4175a50-26b9-459f-9f46-840fdeeb1845"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:14:23 crc kubenswrapper[4710]: I1002 14:14:23.835501 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4175a50-26b9-459f-9f46-840fdeeb1845-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.017435 4710 generic.go:334] "Generic (PLEG): container finished" podID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerID="25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7" exitCode=0 Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.017485 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7cc" event={"ID":"c4175a50-26b9-459f-9f46-840fdeeb1845","Type":"ContainerDied","Data":"25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7"} Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.017517 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7cc" event={"ID":"c4175a50-26b9-459f-9f46-840fdeeb1845","Type":"ContainerDied","Data":"7d87dc1d5fe6c6c72910a5d374eea345c9dd5f5ba000b170c6667c5aefa6a91d"} Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.017523 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7cc" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.017539 4710 scope.go:117] "RemoveContainer" containerID="25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.042054 4710 scope.go:117] "RemoveContainer" containerID="0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.060381 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7cc"] Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.067328 4710 scope.go:117] "RemoveContainer" containerID="6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.070208 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7cc"] Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.128536 4710 scope.go:117] "RemoveContainer" containerID="25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7" Oct 02 14:14:24 crc kubenswrapper[4710]: E1002 14:14:24.129154 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7\": container with ID starting with 25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7 not found: ID does not exist" containerID="25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.129225 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7"} err="failed to get container status \"25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7\": rpc error: code = NotFound desc = could not find container \"25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7\": container with ID starting with 25515c112c914df2e83694f605e5d735ebbb23d9f0de1feccc9098637c2f0dd7 not found: ID does not exist" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.129270 4710 scope.go:117] "RemoveContainer" containerID="0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6" Oct 02 14:14:24 crc kubenswrapper[4710]: E1002 14:14:24.129833 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6\": container with ID starting with 0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6 not found: ID does not exist" containerID="0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.129877 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6"} err="failed to get container status \"0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6\": rpc error: code = NotFound desc = could not find container \"0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6\": container with ID starting with 0cf6d3a7d9acadb903b4c67b1f0790c6abd3d78fc318b67167ff28581585bca6 not found: ID does not exist" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.129906 4710 scope.go:117] "RemoveContainer" containerID="6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8" Oct 02 14:14:24 crc kubenswrapper[4710]: E1002 14:14:24.130276 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8\": container with ID starting with 6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8 not found: ID does not exist" containerID="6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.130314 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8"} err="failed to get container status \"6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8\": rpc error: code = NotFound desc = could not find container \"6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8\": container with ID starting with 6238be662dc943ded891c37f60df22293ef00cbc54b4a2b5a26e0c80673716e8 not found: ID does not exist" Oct 02 14:14:24 crc kubenswrapper[4710]: I1002 14:14:24.915055 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" path="/var/lib/kubelet/pods/c4175a50-26b9-459f-9f46-840fdeeb1845/volumes" Oct 02 14:14:32 crc kubenswrapper[4710]: I1002 14:14:32.905302 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:14:32 crc kubenswrapper[4710]: E1002 14:14:32.906382 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:14:42 crc kubenswrapper[4710]: E1002 14:14:42.854947 4710 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.103:56208->38.102.83.103:33121: write tcp 38.102.83.103:56208->38.102.83.103:33121: write: broken pipe Oct 02 14:14:45 crc kubenswrapper[4710]: E1002 14:14:45.946628 4710 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.103:45162->38.102.83.103:33121: write tcp 38.102.83.103:45162->38.102.83.103:33121: write: broken pipe Oct 02 14:14:47 crc kubenswrapper[4710]: I1002 14:14:47.905499 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:14:47 crc kubenswrapper[4710]: E1002 14:14:47.906181 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.153480 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl"] Oct 02 14:15:00 crc kubenswrapper[4710]: E1002 14:15:00.154480 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="registry-server" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.154498 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="registry-server" Oct 02 14:15:00 crc kubenswrapper[4710]: E1002 14:15:00.154518 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="extract-content" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.154525 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="extract-content" Oct 02 14:15:00 crc kubenswrapper[4710]: E1002 14:15:00.154534 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="extract-utilities" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.154540 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="extract-utilities" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.154735 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4175a50-26b9-459f-9f46-840fdeeb1845" containerName="registry-server" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.155456 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.158113 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.160011 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.181807 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl"] Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.258296 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqv4j\" (UniqueName: \"kubernetes.io/projected/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-kube-api-access-cqv4j\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.259176 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-config-volume\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.259439 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-secret-volume\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.361548 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-config-volume\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.361624 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-secret-volume\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.361774 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqv4j\" (UniqueName: \"kubernetes.io/projected/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-kube-api-access-cqv4j\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.362543 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-config-volume\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.369138 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-secret-volume\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.381715 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqv4j\" (UniqueName: \"kubernetes.io/projected/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-kube-api-access-cqv4j\") pod \"collect-profiles-29323575-xqzrl\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.486213 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.904652 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:15:00 crc kubenswrapper[4710]: E1002 14:15:00.905410 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:15:00 crc kubenswrapper[4710]: I1002 14:15:00.959979 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl"] Oct 02 14:15:01 crc kubenswrapper[4710]: W1002 14:15:01.436863 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcf03ad5_533d_4ff0_b4ac_385bf13d03b2.slice/crio-fdab20c15385898eac5cbdbf727e023f99ccbb6bdf59abe8d5ed17fe1da515b0 WatchSource:0}: Error finding container fdab20c15385898eac5cbdbf727e023f99ccbb6bdf59abe8d5ed17fe1da515b0: Status 404 returned error can't find the container with id fdab20c15385898eac5cbdbf727e023f99ccbb6bdf59abe8d5ed17fe1da515b0 Oct 02 14:15:02 crc kubenswrapper[4710]: I1002 14:15:02.435849 4710 generic.go:334] "Generic (PLEG): container finished" podID="fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" containerID="8aca55b0a7b55d8fcd515b6b33b8000dd7bb2789c81cd41b5ea8961c40b9f077" exitCode=0 Oct 02 14:15:02 crc kubenswrapper[4710]: I1002 14:15:02.435931 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" event={"ID":"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2","Type":"ContainerDied","Data":"8aca55b0a7b55d8fcd515b6b33b8000dd7bb2789c81cd41b5ea8961c40b9f077"} Oct 02 14:15:02 crc kubenswrapper[4710]: I1002 14:15:02.435980 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" event={"ID":"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2","Type":"ContainerStarted","Data":"fdab20c15385898eac5cbdbf727e023f99ccbb6bdf59abe8d5ed17fe1da515b0"} Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.799981 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.853608 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-config-volume\") pod \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.853702 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-secret-volume\") pod \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.853866 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqv4j\" (UniqueName: \"kubernetes.io/projected/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-kube-api-access-cqv4j\") pod \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\" (UID: \"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2\") " Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.854504 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-config-volume" (OuterVolumeSpecName: "config-volume") pod "fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" (UID: "fcf03ad5-533d-4ff0-b4ac-385bf13d03b2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.860642 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" (UID: "fcf03ad5-533d-4ff0-b4ac-385bf13d03b2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.871012 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-kube-api-access-cqv4j" (OuterVolumeSpecName: "kube-api-access-cqv4j") pod "fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" (UID: "fcf03ad5-533d-4ff0-b4ac-385bf13d03b2"). InnerVolumeSpecName "kube-api-access-cqv4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.955833 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.955879 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 14:15:03 crc kubenswrapper[4710]: I1002 14:15:03.955896 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqv4j\" (UniqueName: \"kubernetes.io/projected/fcf03ad5-533d-4ff0-b4ac-385bf13d03b2-kube-api-access-cqv4j\") on node \"crc\" DevicePath \"\"" Oct 02 14:15:04 crc kubenswrapper[4710]: I1002 14:15:04.461861 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" event={"ID":"fcf03ad5-533d-4ff0-b4ac-385bf13d03b2","Type":"ContainerDied","Data":"fdab20c15385898eac5cbdbf727e023f99ccbb6bdf59abe8d5ed17fe1da515b0"} Oct 02 14:15:04 crc kubenswrapper[4710]: I1002 14:15:04.461916 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdab20c15385898eac5cbdbf727e023f99ccbb6bdf59abe8d5ed17fe1da515b0" Oct 02 14:15:04 crc kubenswrapper[4710]: I1002 14:15:04.461912 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323575-xqzrl" Oct 02 14:15:04 crc kubenswrapper[4710]: I1002 14:15:04.889331 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72"] Oct 02 14:15:04 crc kubenswrapper[4710]: I1002 14:15:04.898368 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323530-8zt72"] Oct 02 14:15:04 crc kubenswrapper[4710]: I1002 14:15:04.917071 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f" path="/var/lib/kubelet/pods/e22c73b0-a1a3-4241-a2d3-7b5a1bb0078f/volumes" Oct 02 14:15:12 crc kubenswrapper[4710]: I1002 14:15:12.905806 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:15:12 crc kubenswrapper[4710]: E1002 14:15:12.907069 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:15:24 crc kubenswrapper[4710]: I1002 14:15:24.904862 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:15:24 crc kubenswrapper[4710]: E1002 14:15:24.907049 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:15:37 crc kubenswrapper[4710]: I1002 14:15:37.905032 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:15:37 crc kubenswrapper[4710]: E1002 14:15:37.905693 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:15:48 crc kubenswrapper[4710]: I1002 14:15:48.912494 4710 scope.go:117] "RemoveContainer" containerID="4572e5a2057d55eabe5df7321a3aba9b8e05d18c81c9682c5c81fb53f28cad65" Oct 02 14:15:49 crc kubenswrapper[4710]: I1002 14:15:49.905341 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:15:49 crc kubenswrapper[4710]: E1002 14:15:49.906231 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:16:04 crc kubenswrapper[4710]: I1002 14:16:04.904094 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:16:04 crc kubenswrapper[4710]: E1002 14:16:04.905959 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:16:17 crc kubenswrapper[4710]: I1002 14:16:17.904065 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:16:17 crc kubenswrapper[4710]: E1002 14:16:17.904818 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:16:31 crc kubenswrapper[4710]: I1002 14:16:31.904654 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:16:31 crc kubenswrapper[4710]: E1002 14:16:31.905279 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:16:42 crc kubenswrapper[4710]: I1002 14:16:42.905380 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:16:42 crc kubenswrapper[4710]: E1002 14:16:42.906725 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.100445 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l5k29"] Oct 02 14:16:51 crc kubenswrapper[4710]: E1002 14:16:51.101779 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" containerName="collect-profiles" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.101798 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" containerName="collect-profiles" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.101997 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf03ad5-533d-4ff0-b4ac-385bf13d03b2" containerName="collect-profiles" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.105925 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.121957 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5k29"] Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.250765 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-utilities\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.251260 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h529x\" (UniqueName: \"kubernetes.io/projected/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-kube-api-access-h529x\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.251287 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.353324 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h529x\" (UniqueName: \"kubernetes.io/projected/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-kube-api-access-h529x\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.353376 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.353445 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-utilities\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.354095 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.354162 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-utilities\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.380638 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h529x\" (UniqueName: \"kubernetes.io/projected/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-kube-api-access-h529x\") pod \"redhat-operators-l5k29\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:51 crc kubenswrapper[4710]: I1002 14:16:51.495548 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:16:52 crc kubenswrapper[4710]: I1002 14:16:51.999786 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5k29"] Oct 02 14:16:52 crc kubenswrapper[4710]: I1002 14:16:52.680738 4710 generic.go:334] "Generic (PLEG): container finished" podID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerID="f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9" exitCode=0 Oct 02 14:16:52 crc kubenswrapper[4710]: I1002 14:16:52.680825 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerDied","Data":"f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9"} Oct 02 14:16:52 crc kubenswrapper[4710]: I1002 14:16:52.681294 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerStarted","Data":"363b7988b02c3ac5982b65b2c4f422ccc707165cd21f20a3a089a1f5ff8caa7b"} Oct 02 14:16:54 crc kubenswrapper[4710]: I1002 14:16:54.904982 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:16:54 crc kubenswrapper[4710]: E1002 14:16:54.905864 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:16:56 crc kubenswrapper[4710]: I1002 14:16:56.725823 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerStarted","Data":"b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0"} Oct 02 14:16:58 crc kubenswrapper[4710]: I1002 14:16:58.757770 4710 generic.go:334] "Generic (PLEG): container finished" podID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerID="b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0" exitCode=0 Oct 02 14:16:58 crc kubenswrapper[4710]: I1002 14:16:58.757862 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerDied","Data":"b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0"} Oct 02 14:16:59 crc kubenswrapper[4710]: I1002 14:16:59.769365 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerStarted","Data":"46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0"} Oct 02 14:16:59 crc kubenswrapper[4710]: I1002 14:16:59.799310 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l5k29" podStartSLOduration=2.2430119 podStartE2EDuration="8.799283502s" podCreationTimestamp="2025-10-02 14:16:51 +0000 UTC" firstStartedPulling="2025-10-02 14:16:52.683412837 +0000 UTC m=+5256.789823760" lastFinishedPulling="2025-10-02 14:16:59.239684469 +0000 UTC m=+5263.346095362" observedRunningTime="2025-10-02 14:16:59.797493597 +0000 UTC m=+5263.903904480" watchObservedRunningTime="2025-10-02 14:16:59.799283502 +0000 UTC m=+5263.905694425" Oct 02 14:17:01 crc kubenswrapper[4710]: I1002 14:17:01.496427 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:17:01 crc kubenswrapper[4710]: I1002 14:17:01.497143 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:17:02 crc kubenswrapper[4710]: I1002 14:17:02.578141 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l5k29" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="registry-server" probeResult="failure" output=< Oct 02 14:17:02 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 14:17:02 crc kubenswrapper[4710]: > Oct 02 14:17:06 crc kubenswrapper[4710]: I1002 14:17:06.914883 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:17:06 crc kubenswrapper[4710]: E1002 14:17:06.915356 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:17:11 crc kubenswrapper[4710]: I1002 14:17:11.572911 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:17:11 crc kubenswrapper[4710]: I1002 14:17:11.645869 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:17:11 crc kubenswrapper[4710]: I1002 14:17:11.825294 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5k29"] Oct 02 14:17:12 crc kubenswrapper[4710]: I1002 14:17:12.922718 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l5k29" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="registry-server" containerID="cri-o://46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0" gracePeriod=2 Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.466923 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.571596 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h529x\" (UniqueName: \"kubernetes.io/projected/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-kube-api-access-h529x\") pod \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.571789 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-utilities\") pod \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.571861 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content\") pod \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.572628 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-utilities" (OuterVolumeSpecName: "utilities") pod "c90b5ced-5c5e-4877-9870-a1e2be6e55d6" (UID: "c90b5ced-5c5e-4877-9870-a1e2be6e55d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.577763 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-kube-api-access-h529x" (OuterVolumeSpecName: "kube-api-access-h529x") pod "c90b5ced-5c5e-4877-9870-a1e2be6e55d6" (UID: "c90b5ced-5c5e-4877-9870-a1e2be6e55d6"). InnerVolumeSpecName "kube-api-access-h529x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.672540 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c90b5ced-5c5e-4877-9870-a1e2be6e55d6" (UID: "c90b5ced-5c5e-4877-9870-a1e2be6e55d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.673116 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content\") pod \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\" (UID: \"c90b5ced-5c5e-4877-9870-a1e2be6e55d6\") " Oct 02 14:17:13 crc kubenswrapper[4710]: W1002 14:17:13.673722 4710 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c90b5ced-5c5e-4877-9870-a1e2be6e55d6/volumes/kubernetes.io~empty-dir/catalog-content Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.673781 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c90b5ced-5c5e-4877-9870-a1e2be6e55d6" (UID: "c90b5ced-5c5e-4877-9870-a1e2be6e55d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.673824 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h529x\" (UniqueName: \"kubernetes.io/projected/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-kube-api-access-h529x\") on node \"crc\" DevicePath \"\"" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.673838 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.673850 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b5ced-5c5e-4877-9870-a1e2be6e55d6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.935552 4710 generic.go:334] "Generic (PLEG): container finished" podID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerID="46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0" exitCode=0 Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.935682 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerDied","Data":"46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0"} Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.936109 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5k29" event={"ID":"c90b5ced-5c5e-4877-9870-a1e2be6e55d6","Type":"ContainerDied","Data":"363b7988b02c3ac5982b65b2c4f422ccc707165cd21f20a3a089a1f5ff8caa7b"} Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.936157 4710 scope.go:117] "RemoveContainer" containerID="46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.935697 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5k29" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.962030 4710 scope.go:117] "RemoveContainer" containerID="b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0" Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.980371 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5k29"] Oct 02 14:17:13 crc kubenswrapper[4710]: I1002 14:17:13.989974 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l5k29"] Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.007286 4710 scope.go:117] "RemoveContainer" containerID="f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.056780 4710 scope.go:117] "RemoveContainer" containerID="46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0" Oct 02 14:17:14 crc kubenswrapper[4710]: E1002 14:17:14.060391 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0\": container with ID starting with 46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0 not found: ID does not exist" containerID="46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.060435 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0"} err="failed to get container status \"46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0\": rpc error: code = NotFound desc = could not find container \"46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0\": container with ID starting with 46a793a5e75f6c723fa2c24ddbb67e57c03bd590e4880b04f6f80f43d5f93ab0 not found: ID does not exist" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.060462 4710 scope.go:117] "RemoveContainer" containerID="b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0" Oct 02 14:17:14 crc kubenswrapper[4710]: E1002 14:17:14.060819 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0\": container with ID starting with b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0 not found: ID does not exist" containerID="b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.060857 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0"} err="failed to get container status \"b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0\": rpc error: code = NotFound desc = could not find container \"b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0\": container with ID starting with b50a341297855e64a39520b4525c160a1260748c84ac33839ca4173932b4e2f0 not found: ID does not exist" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.060886 4710 scope.go:117] "RemoveContainer" containerID="f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9" Oct 02 14:17:14 crc kubenswrapper[4710]: E1002 14:17:14.061343 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9\": container with ID starting with f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9 not found: ID does not exist" containerID="f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.061382 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9"} err="failed to get container status \"f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9\": rpc error: code = NotFound desc = could not find container \"f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9\": container with ID starting with f4e8601202fc29abd5c3bc795e688bdd2bde331c1ac7791070b1085df9f308d9 not found: ID does not exist" Oct 02 14:17:14 crc kubenswrapper[4710]: I1002 14:17:14.927274 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" path="/var/lib/kubelet/pods/c90b5ced-5c5e-4877-9870-a1e2be6e55d6/volumes" Oct 02 14:17:21 crc kubenswrapper[4710]: I1002 14:17:21.904810 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:17:21 crc kubenswrapper[4710]: E1002 14:17:21.906093 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:17:36 crc kubenswrapper[4710]: I1002 14:17:36.912617 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:17:36 crc kubenswrapper[4710]: E1002 14:17:36.913229 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:17:49 crc kubenswrapper[4710]: I1002 14:17:49.904434 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:17:49 crc kubenswrapper[4710]: E1002 14:17:49.905025 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:18:01 crc kubenswrapper[4710]: I1002 14:18:01.904937 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:18:01 crc kubenswrapper[4710]: E1002 14:18:01.905692 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:18:14 crc kubenswrapper[4710]: I1002 14:18:14.905126 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:18:14 crc kubenswrapper[4710]: E1002 14:18:14.905880 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:18:28 crc kubenswrapper[4710]: I1002 14:18:28.905002 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:18:28 crc kubenswrapper[4710]: E1002 14:18:28.905867 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:18:40 crc kubenswrapper[4710]: I1002 14:18:40.905795 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:18:40 crc kubenswrapper[4710]: E1002 14:18:40.907206 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:18:53 crc kubenswrapper[4710]: I1002 14:18:53.904528 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:18:55 crc kubenswrapper[4710]: I1002 14:18:55.069633 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"5ac9ea6b4e21821de52736c4c61c86ba05c2a3f9ecbd59da65453c4a85595777"} Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.705585 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4w2p7"] Oct 02 14:19:54 crc kubenswrapper[4710]: E1002 14:19:54.707505 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="extract-content" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.707536 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="extract-content" Oct 02 14:19:54 crc kubenswrapper[4710]: E1002 14:19:54.707611 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="extract-utilities" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.707632 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="extract-utilities" Oct 02 14:19:54 crc kubenswrapper[4710]: E1002 14:19:54.707671 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="registry-server" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.707686 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="registry-server" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.708107 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="c90b5ced-5c5e-4877-9870-a1e2be6e55d6" containerName="registry-server" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.711678 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.720237 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4w2p7"] Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.874716 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-catalog-content\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.874884 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnzzm\" (UniqueName: \"kubernetes.io/projected/abb0dc7a-3444-4b6a-a19c-1dc42695758c-kube-api-access-qnzzm\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.875047 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-utilities\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.976972 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-catalog-content\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.977045 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnzzm\" (UniqueName: \"kubernetes.io/projected/abb0dc7a-3444-4b6a-a19c-1dc42695758c-kube-api-access-qnzzm\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.977119 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-utilities\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.977670 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-utilities\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:54 crc kubenswrapper[4710]: I1002 14:19:54.978033 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-catalog-content\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:55 crc kubenswrapper[4710]: I1002 14:19:55.004896 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnzzm\" (UniqueName: \"kubernetes.io/projected/abb0dc7a-3444-4b6a-a19c-1dc42695758c-kube-api-access-qnzzm\") pod \"certified-operators-4w2p7\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:55 crc kubenswrapper[4710]: I1002 14:19:55.041217 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:19:55 crc kubenswrapper[4710]: I1002 14:19:55.574950 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4w2p7"] Oct 02 14:19:55 crc kubenswrapper[4710]: I1002 14:19:55.778672 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4w2p7" event={"ID":"abb0dc7a-3444-4b6a-a19c-1dc42695758c","Type":"ContainerStarted","Data":"ddcc98fccd0c177b761bc94fe6d5a50ccf348b2c0aeb7862cfce556329040ed2"} Oct 02 14:19:56 crc kubenswrapper[4710]: I1002 14:19:56.791413 4710 generic.go:334] "Generic (PLEG): container finished" podID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerID="780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4" exitCode=0 Oct 02 14:19:56 crc kubenswrapper[4710]: I1002 14:19:56.791499 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4w2p7" event={"ID":"abb0dc7a-3444-4b6a-a19c-1dc42695758c","Type":"ContainerDied","Data":"780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4"} Oct 02 14:19:56 crc kubenswrapper[4710]: I1002 14:19:56.793448 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 14:19:58 crc kubenswrapper[4710]: E1002 14:19:58.414515 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabb0dc7a_3444_4b6a_a19c_1dc42695758c.slice/crio-4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabb0dc7a_3444_4b6a_a19c_1dc42695758c.slice/crio-conmon-4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90.scope\": RecentStats: unable to find data in memory cache]" Oct 02 14:19:58 crc kubenswrapper[4710]: I1002 14:19:58.815679 4710 generic.go:334] "Generic (PLEG): container finished" podID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerID="4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90" exitCode=0 Oct 02 14:19:58 crc kubenswrapper[4710]: I1002 14:19:58.815730 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4w2p7" event={"ID":"abb0dc7a-3444-4b6a-a19c-1dc42695758c","Type":"ContainerDied","Data":"4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90"} Oct 02 14:19:59 crc kubenswrapper[4710]: I1002 14:19:59.827889 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4w2p7" event={"ID":"abb0dc7a-3444-4b6a-a19c-1dc42695758c","Type":"ContainerStarted","Data":"57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19"} Oct 02 14:19:59 crc kubenswrapper[4710]: I1002 14:19:59.854020 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4w2p7" podStartSLOduration=3.415909982 podStartE2EDuration="5.853994462s" podCreationTimestamp="2025-10-02 14:19:54 +0000 UTC" firstStartedPulling="2025-10-02 14:19:56.793201289 +0000 UTC m=+5440.899612172" lastFinishedPulling="2025-10-02 14:19:59.231285749 +0000 UTC m=+5443.337696652" observedRunningTime="2025-10-02 14:19:59.845269662 +0000 UTC m=+5443.951680555" watchObservedRunningTime="2025-10-02 14:19:59.853994462 +0000 UTC m=+5443.960405355" Oct 02 14:20:05 crc kubenswrapper[4710]: I1002 14:20:05.042182 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:20:05 crc kubenswrapper[4710]: I1002 14:20:05.043101 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:20:05 crc kubenswrapper[4710]: I1002 14:20:05.142704 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:20:06 crc kubenswrapper[4710]: I1002 14:20:06.406322 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:20:06 crc kubenswrapper[4710]: I1002 14:20:06.471218 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4w2p7"] Oct 02 14:20:07 crc kubenswrapper[4710]: I1002 14:20:07.927495 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4w2p7" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="registry-server" containerID="cri-o://57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19" gracePeriod=2 Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.516797 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.581940 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnzzm\" (UniqueName: \"kubernetes.io/projected/abb0dc7a-3444-4b6a-a19c-1dc42695758c-kube-api-access-qnzzm\") pod \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.582052 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-utilities\") pod \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.582118 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-catalog-content\") pod \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\" (UID: \"abb0dc7a-3444-4b6a-a19c-1dc42695758c\") " Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.583027 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-utilities" (OuterVolumeSpecName: "utilities") pod "abb0dc7a-3444-4b6a-a19c-1dc42695758c" (UID: "abb0dc7a-3444-4b6a-a19c-1dc42695758c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.588134 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb0dc7a-3444-4b6a-a19c-1dc42695758c-kube-api-access-qnzzm" (OuterVolumeSpecName: "kube-api-access-qnzzm") pod "abb0dc7a-3444-4b6a-a19c-1dc42695758c" (UID: "abb0dc7a-3444-4b6a-a19c-1dc42695758c"). InnerVolumeSpecName "kube-api-access-qnzzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.624536 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abb0dc7a-3444-4b6a-a19c-1dc42695758c" (UID: "abb0dc7a-3444-4b6a-a19c-1dc42695758c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.685237 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnzzm\" (UniqueName: \"kubernetes.io/projected/abb0dc7a-3444-4b6a-a19c-1dc42695758c-kube-api-access-qnzzm\") on node \"crc\" DevicePath \"\"" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.685304 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.685332 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb0dc7a-3444-4b6a-a19c-1dc42695758c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.941098 4710 generic.go:334] "Generic (PLEG): container finished" podID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerID="57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19" exitCode=0 Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.941150 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4w2p7" event={"ID":"abb0dc7a-3444-4b6a-a19c-1dc42695758c","Type":"ContainerDied","Data":"57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19"} Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.941182 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4w2p7" event={"ID":"abb0dc7a-3444-4b6a-a19c-1dc42695758c","Type":"ContainerDied","Data":"ddcc98fccd0c177b761bc94fe6d5a50ccf348b2c0aeb7862cfce556329040ed2"} Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.941202 4710 scope.go:117] "RemoveContainer" containerID="57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.941234 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4w2p7" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.977910 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4w2p7"] Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.985162 4710 scope.go:117] "RemoveContainer" containerID="4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90" Oct 02 14:20:08 crc kubenswrapper[4710]: I1002 14:20:08.989051 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4w2p7"] Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.025181 4710 scope.go:117] "RemoveContainer" containerID="780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4" Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.066170 4710 scope.go:117] "RemoveContainer" containerID="57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19" Oct 02 14:20:09 crc kubenswrapper[4710]: E1002 14:20:09.066605 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19\": container with ID starting with 57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19 not found: ID does not exist" containerID="57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19" Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.066634 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19"} err="failed to get container status \"57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19\": rpc error: code = NotFound desc = could not find container \"57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19\": container with ID starting with 57207d93723b1950e0b22dc4029880bae7bf8b4ddb181f7192508cc6e4f10e19 not found: ID does not exist" Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.066653 4710 scope.go:117] "RemoveContainer" containerID="4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90" Oct 02 14:20:09 crc kubenswrapper[4710]: E1002 14:20:09.067112 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90\": container with ID starting with 4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90 not found: ID does not exist" containerID="4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90" Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.067133 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90"} err="failed to get container status \"4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90\": rpc error: code = NotFound desc = could not find container \"4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90\": container with ID starting with 4f35fee3fb0ba26ea0560f7cfb3f2436b70ba49e4c71241eebcc6809c131bb90 not found: ID does not exist" Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.067145 4710 scope.go:117] "RemoveContainer" containerID="780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4" Oct 02 14:20:09 crc kubenswrapper[4710]: E1002 14:20:09.067474 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4\": container with ID starting with 780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4 not found: ID does not exist" containerID="780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4" Oct 02 14:20:09 crc kubenswrapper[4710]: I1002 14:20:09.067505 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4"} err="failed to get container status \"780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4\": rpc error: code = NotFound desc = could not find container \"780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4\": container with ID starting with 780c7f88ca55ded5b5ef06c67f5228d700a3e8728a38e02eab08d7f38946d6b4 not found: ID does not exist" Oct 02 14:20:10 crc kubenswrapper[4710]: I1002 14:20:10.922513 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" path="/var/lib/kubelet/pods/abb0dc7a-3444-4b6a-a19c-1dc42695758c/volumes" Oct 02 14:21:22 crc kubenswrapper[4710]: I1002 14:21:22.530106 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:21:22 crc kubenswrapper[4710]: I1002 14:21:22.530586 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:21:52 crc kubenswrapper[4710]: I1002 14:21:52.530699 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:21:52 crc kubenswrapper[4710]: I1002 14:21:52.531137 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:22:07 crc kubenswrapper[4710]: I1002 14:22:07.996793 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kxbkr"] Oct 02 14:22:07 crc kubenswrapper[4710]: E1002 14:22:07.998641 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="registry-server" Oct 02 14:22:07 crc kubenswrapper[4710]: I1002 14:22:07.998719 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="registry-server" Oct 02 14:22:07 crc kubenswrapper[4710]: E1002 14:22:07.998816 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="extract-utilities" Oct 02 14:22:07 crc kubenswrapper[4710]: I1002 14:22:07.998874 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="extract-utilities" Oct 02 14:22:07 crc kubenswrapper[4710]: E1002 14:22:07.998940 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="extract-content" Oct 02 14:22:07 crc kubenswrapper[4710]: I1002 14:22:07.999008 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="extract-content" Oct 02 14:22:07 crc kubenswrapper[4710]: I1002 14:22:07.999257 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb0dc7a-3444-4b6a-a19c-1dc42695758c" containerName="registry-server" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.001701 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.029546 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxbkr"] Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.095350 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-utilities\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.095409 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-catalog-content\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.095516 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdx87\" (UniqueName: \"kubernetes.io/projected/66659b02-f45c-408f-8fe2-74ce059d36f3-kube-api-access-cdx87\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.197246 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-utilities\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.197319 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-catalog-content\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.197456 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdx87\" (UniqueName: \"kubernetes.io/projected/66659b02-f45c-408f-8fe2-74ce059d36f3-kube-api-access-cdx87\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.197953 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-utilities\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.197976 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-catalog-content\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.221041 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdx87\" (UniqueName: \"kubernetes.io/projected/66659b02-f45c-408f-8fe2-74ce059d36f3-kube-api-access-cdx87\") pod \"community-operators-kxbkr\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.386208 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:08 crc kubenswrapper[4710]: I1002 14:22:08.901502 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kxbkr"] Oct 02 14:22:09 crc kubenswrapper[4710]: I1002 14:22:09.292414 4710 generic.go:334] "Generic (PLEG): container finished" podID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerID="417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107" exitCode=0 Oct 02 14:22:09 crc kubenswrapper[4710]: I1002 14:22:09.292840 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerDied","Data":"417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107"} Oct 02 14:22:09 crc kubenswrapper[4710]: I1002 14:22:09.292916 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerStarted","Data":"0be9e94abfdad51224243c3c98bce4010fccd8b15f9608a176ad6c28f4ae447b"} Oct 02 14:22:10 crc kubenswrapper[4710]: I1002 14:22:10.306667 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerStarted","Data":"1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3"} Oct 02 14:22:11 crc kubenswrapper[4710]: I1002 14:22:11.320629 4710 generic.go:334] "Generic (PLEG): container finished" podID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerID="1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3" exitCode=0 Oct 02 14:22:11 crc kubenswrapper[4710]: I1002 14:22:11.320686 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerDied","Data":"1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3"} Oct 02 14:22:12 crc kubenswrapper[4710]: I1002 14:22:12.341372 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerStarted","Data":"dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67"} Oct 02 14:22:12 crc kubenswrapper[4710]: I1002 14:22:12.359321 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kxbkr" podStartSLOduration=2.895082803 podStartE2EDuration="5.359301682s" podCreationTimestamp="2025-10-02 14:22:07 +0000 UTC" firstStartedPulling="2025-10-02 14:22:09.296026006 +0000 UTC m=+5573.402436889" lastFinishedPulling="2025-10-02 14:22:11.760244845 +0000 UTC m=+5575.866655768" observedRunningTime="2025-10-02 14:22:12.357616949 +0000 UTC m=+5576.464027892" watchObservedRunningTime="2025-10-02 14:22:12.359301682 +0000 UTC m=+5576.465712555" Oct 02 14:22:18 crc kubenswrapper[4710]: I1002 14:22:18.387488 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:18 crc kubenswrapper[4710]: I1002 14:22:18.388150 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:18 crc kubenswrapper[4710]: I1002 14:22:18.442368 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:18 crc kubenswrapper[4710]: I1002 14:22:18.555588 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:18 crc kubenswrapper[4710]: I1002 14:22:18.684703 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxbkr"] Oct 02 14:22:20 crc kubenswrapper[4710]: I1002 14:22:20.466590 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kxbkr" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="registry-server" containerID="cri-o://dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67" gracePeriod=2 Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.027116 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.215958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-catalog-content\") pod \"66659b02-f45c-408f-8fe2-74ce059d36f3\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.216086 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdx87\" (UniqueName: \"kubernetes.io/projected/66659b02-f45c-408f-8fe2-74ce059d36f3-kube-api-access-cdx87\") pod \"66659b02-f45c-408f-8fe2-74ce059d36f3\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.216304 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-utilities\") pod \"66659b02-f45c-408f-8fe2-74ce059d36f3\" (UID: \"66659b02-f45c-408f-8fe2-74ce059d36f3\") " Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.217214 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-utilities" (OuterVolumeSpecName: "utilities") pod "66659b02-f45c-408f-8fe2-74ce059d36f3" (UID: "66659b02-f45c-408f-8fe2-74ce059d36f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.229032 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66659b02-f45c-408f-8fe2-74ce059d36f3-kube-api-access-cdx87" (OuterVolumeSpecName: "kube-api-access-cdx87") pod "66659b02-f45c-408f-8fe2-74ce059d36f3" (UID: "66659b02-f45c-408f-8fe2-74ce059d36f3"). InnerVolumeSpecName "kube-api-access-cdx87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.272836 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66659b02-f45c-408f-8fe2-74ce059d36f3" (UID: "66659b02-f45c-408f-8fe2-74ce059d36f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.318323 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.318371 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66659b02-f45c-408f-8fe2-74ce059d36f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.318399 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdx87\" (UniqueName: \"kubernetes.io/projected/66659b02-f45c-408f-8fe2-74ce059d36f3-kube-api-access-cdx87\") on node \"crc\" DevicePath \"\"" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.479040 4710 generic.go:334] "Generic (PLEG): container finished" podID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerID="dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67" exitCode=0 Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.479127 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kxbkr" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.479141 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerDied","Data":"dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67"} Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.479251 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kxbkr" event={"ID":"66659b02-f45c-408f-8fe2-74ce059d36f3","Type":"ContainerDied","Data":"0be9e94abfdad51224243c3c98bce4010fccd8b15f9608a176ad6c28f4ae447b"} Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.479289 4710 scope.go:117] "RemoveContainer" containerID="dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.498280 4710 scope.go:117] "RemoveContainer" containerID="1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.527422 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kxbkr"] Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.531060 4710 scope.go:117] "RemoveContainer" containerID="417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.557354 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kxbkr"] Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.581835 4710 scope.go:117] "RemoveContainer" containerID="dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67" Oct 02 14:22:21 crc kubenswrapper[4710]: E1002 14:22:21.582245 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67\": container with ID starting with dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67 not found: ID does not exist" containerID="dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.582298 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67"} err="failed to get container status \"dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67\": rpc error: code = NotFound desc = could not find container \"dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67\": container with ID starting with dd511ded163818d6e012a0ee4b63f85b0d48ffbd1c27ba0f0ffe282ebce0fe67 not found: ID does not exist" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.582334 4710 scope.go:117] "RemoveContainer" containerID="1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3" Oct 02 14:22:21 crc kubenswrapper[4710]: E1002 14:22:21.583011 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3\": container with ID starting with 1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3 not found: ID does not exist" containerID="1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.583056 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3"} err="failed to get container status \"1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3\": rpc error: code = NotFound desc = could not find container \"1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3\": container with ID starting with 1b017291a50f4cdbde11e2eb90f2fa30e1a3f5e6cb29cc2f82c89ec73fce82f3 not found: ID does not exist" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.583089 4710 scope.go:117] "RemoveContainer" containerID="417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107" Oct 02 14:22:21 crc kubenswrapper[4710]: E1002 14:22:21.583548 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107\": container with ID starting with 417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107 not found: ID does not exist" containerID="417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107" Oct 02 14:22:21 crc kubenswrapper[4710]: I1002 14:22:21.583586 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107"} err="failed to get container status \"417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107\": rpc error: code = NotFound desc = could not find container \"417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107\": container with ID starting with 417922cacab6d67af746e07318d10b785a535b9f79c01cf77693ec1a2035b107 not found: ID does not exist" Oct 02 14:22:22 crc kubenswrapper[4710]: I1002 14:22:22.530576 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:22:22 crc kubenswrapper[4710]: I1002 14:22:22.530964 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:22:22 crc kubenswrapper[4710]: I1002 14:22:22.531019 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 14:22:22 crc kubenswrapper[4710]: I1002 14:22:22.531916 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ac9ea6b4e21821de52736c4c61c86ba05c2a3f9ecbd59da65453c4a85595777"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 14:22:22 crc kubenswrapper[4710]: I1002 14:22:22.531969 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://5ac9ea6b4e21821de52736c4c61c86ba05c2a3f9ecbd59da65453c4a85595777" gracePeriod=600 Oct 02 14:22:22 crc kubenswrapper[4710]: I1002 14:22:22.914825 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" path="/var/lib/kubelet/pods/66659b02-f45c-408f-8fe2-74ce059d36f3/volumes" Oct 02 14:22:23 crc kubenswrapper[4710]: I1002 14:22:23.501431 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="5ac9ea6b4e21821de52736c4c61c86ba05c2a3f9ecbd59da65453c4a85595777" exitCode=0 Oct 02 14:22:23 crc kubenswrapper[4710]: I1002 14:22:23.502122 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"5ac9ea6b4e21821de52736c4c61c86ba05c2a3f9ecbd59da65453c4a85595777"} Oct 02 14:22:23 crc kubenswrapper[4710]: I1002 14:22:23.502244 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a"} Oct 02 14:22:23 crc kubenswrapper[4710]: I1002 14:22:23.502366 4710 scope.go:117] "RemoveContainer" containerID="137634c08e0d8a48d89f99832fe0fc54cd3a0ad04fd64d8871b534696157bb10" Oct 02 14:24:22 crc kubenswrapper[4710]: I1002 14:24:22.530955 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:24:22 crc kubenswrapper[4710]: I1002 14:24:22.531677 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:24:27 crc kubenswrapper[4710]: I1002 14:24:27.951032 4710 generic.go:334] "Generic (PLEG): container finished" podID="22910690-8789-41f2-bd0a-b56b2c6683a6" containerID="66439e1c1481f3a465c44d5c66708295824bc9dea2c8ec2d4e218a74474708e8" exitCode=1 Oct 02 14:24:27 crc kubenswrapper[4710]: I1002 14:24:27.951192 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"22910690-8789-41f2-bd0a-b56b2c6683a6","Type":"ContainerDied","Data":"66439e1c1481f3a465c44d5c66708295824bc9dea2c8ec2d4e218a74474708e8"} Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.364213 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.482880 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28gwc\" (UniqueName: \"kubernetes.io/projected/22910690-8789-41f2-bd0a-b56b2c6683a6-kube-api-access-28gwc\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483013 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ca-certs\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483043 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ssh-key\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483107 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-workdir\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483244 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483287 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-config-data\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483322 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-temporary\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483373 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.483438 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config-secret\") pod \"22910690-8789-41f2-bd0a-b56b2c6683a6\" (UID: \"22910690-8789-41f2-bd0a-b56b2c6683a6\") " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.484549 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.484877 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-config-data" (OuterVolumeSpecName: "config-data") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.488733 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22910690-8789-41f2-bd0a-b56b2c6683a6-kube-api-access-28gwc" (OuterVolumeSpecName: "kube-api-access-28gwc") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "kube-api-access-28gwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.488779 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.488851 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.516091 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.521025 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.533725 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.542043 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "22910690-8789-41f2-bd0a-b56b2c6683a6" (UID: "22910690-8789-41f2-bd0a-b56b2c6683a6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585721 4710 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585813 4710 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585830 4710 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585842 4710 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/22910690-8789-41f2-bd0a-b56b2c6683a6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585854 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585867 4710 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585881 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28gwc\" (UniqueName: \"kubernetes.io/projected/22910690-8789-41f2-bd0a-b56b2c6683a6-kube-api-access-28gwc\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585894 4710 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.585904 4710 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22910690-8789-41f2-bd0a-b56b2c6683a6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.609002 4710 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.687194 4710 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.989440 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"22910690-8789-41f2-bd0a-b56b2c6683a6","Type":"ContainerDied","Data":"2949348ca1c21d04b0748975a3208024671a98e42dcfb287ae00a778a3847734"} Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.989472 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 14:24:29 crc kubenswrapper[4710]: I1002 14:24:29.989483 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2949348ca1c21d04b0748975a3208024671a98e42dcfb287ae00a778a3847734" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.332023 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-25bk6"] Oct 02 14:24:31 crc kubenswrapper[4710]: E1002 14:24:31.333338 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="extract-content" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.333371 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="extract-content" Oct 02 14:24:31 crc kubenswrapper[4710]: E1002 14:24:31.333468 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="extract-utilities" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.333489 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="extract-utilities" Oct 02 14:24:31 crc kubenswrapper[4710]: E1002 14:24:31.333513 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22910690-8789-41f2-bd0a-b56b2c6683a6" containerName="tempest-tests-tempest-tests-runner" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.333534 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="22910690-8789-41f2-bd0a-b56b2c6683a6" containerName="tempest-tests-tempest-tests-runner" Oct 02 14:24:31 crc kubenswrapper[4710]: E1002 14:24:31.333572 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="registry-server" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.333590 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="registry-server" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.339591 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="22910690-8789-41f2-bd0a-b56b2c6683a6" containerName="tempest-tests-tempest-tests-runner" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.339715 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="66659b02-f45c-408f-8fe2-74ce059d36f3" containerName="registry-server" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.343602 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.377265 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25bk6"] Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.421898 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-catalog-content\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.422294 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-utilities\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.422484 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs295\" (UniqueName: \"kubernetes.io/projected/08be0527-86e2-47bb-ad8c-189e51c3e49a-kube-api-access-gs295\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.524791 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-utilities\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.525086 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs295\" (UniqueName: \"kubernetes.io/projected/08be0527-86e2-47bb-ad8c-189e51c3e49a-kube-api-access-gs295\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.525242 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-catalog-content\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.525562 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-utilities\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.525654 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-catalog-content\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.552572 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs295\" (UniqueName: \"kubernetes.io/projected/08be0527-86e2-47bb-ad8c-189e51c3e49a-kube-api-access-gs295\") pod \"redhat-marketplace-25bk6\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:31 crc kubenswrapper[4710]: I1002 14:24:31.685081 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:32 crc kubenswrapper[4710]: I1002 14:24:32.131403 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25bk6"] Oct 02 14:24:33 crc kubenswrapper[4710]: I1002 14:24:33.021684 4710 generic.go:334] "Generic (PLEG): container finished" podID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerID="d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69" exitCode=0 Oct 02 14:24:33 crc kubenswrapper[4710]: I1002 14:24:33.021788 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25bk6" event={"ID":"08be0527-86e2-47bb-ad8c-189e51c3e49a","Type":"ContainerDied","Data":"d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69"} Oct 02 14:24:33 crc kubenswrapper[4710]: I1002 14:24:33.022274 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25bk6" event={"ID":"08be0527-86e2-47bb-ad8c-189e51c3e49a","Type":"ContainerStarted","Data":"2d85bc0a50bf66bef0558cac5c174592bfd22e3cafa2b582723ac2d7241f2beb"} Oct 02 14:24:34 crc kubenswrapper[4710]: I1002 14:24:34.036592 4710 generic.go:334] "Generic (PLEG): container finished" podID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerID="6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3" exitCode=0 Oct 02 14:24:34 crc kubenswrapper[4710]: I1002 14:24:34.036677 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25bk6" event={"ID":"08be0527-86e2-47bb-ad8c-189e51c3e49a","Type":"ContainerDied","Data":"6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3"} Oct 02 14:24:35 crc kubenswrapper[4710]: I1002 14:24:35.068090 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25bk6" event={"ID":"08be0527-86e2-47bb-ad8c-189e51c3e49a","Type":"ContainerStarted","Data":"bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe"} Oct 02 14:24:35 crc kubenswrapper[4710]: I1002 14:24:35.103330 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-25bk6" podStartSLOduration=2.578141173 podStartE2EDuration="4.103313457s" podCreationTimestamp="2025-10-02 14:24:31 +0000 UTC" firstStartedPulling="2025-10-02 14:24:33.023330951 +0000 UTC m=+5717.129741834" lastFinishedPulling="2025-10-02 14:24:34.548503195 +0000 UTC m=+5718.654914118" observedRunningTime="2025-10-02 14:24:35.100153258 +0000 UTC m=+5719.206564181" watchObservedRunningTime="2025-10-02 14:24:35.103313457 +0000 UTC m=+5719.209724340" Oct 02 14:24:36 crc kubenswrapper[4710]: I1002 14:24:36.948423 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 14:24:36 crc kubenswrapper[4710]: I1002 14:24:36.951340 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:36 crc kubenswrapper[4710]: I1002 14:24:36.958209 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pbnjd" Oct 02 14:24:36 crc kubenswrapper[4710]: I1002 14:24:36.960346 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.072314 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.072900 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l8tr\" (UniqueName: \"kubernetes.io/projected/49a43af4-9bec-4b33-a75a-77f0c9a26788-kube-api-access-9l8tr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.174224 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.174391 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l8tr\" (UniqueName: \"kubernetes.io/projected/49a43af4-9bec-4b33-a75a-77f0c9a26788-kube-api-access-9l8tr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.175117 4710 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.208819 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l8tr\" (UniqueName: \"kubernetes.io/projected/49a43af4-9bec-4b33-a75a-77f0c9a26788-kube-api-access-9l8tr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.237635 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49a43af4-9bec-4b33-a75a-77f0c9a26788\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.279131 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 14:24:37 crc kubenswrapper[4710]: I1002 14:24:37.769331 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 14:24:38 crc kubenswrapper[4710]: I1002 14:24:38.103382 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"49a43af4-9bec-4b33-a75a-77f0c9a26788","Type":"ContainerStarted","Data":"3a0949aad918586954bdb8079c16bea7ab0806eea1081e7771656516b95aa756"} Oct 02 14:24:39 crc kubenswrapper[4710]: I1002 14:24:39.118272 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"49a43af4-9bec-4b33-a75a-77f0c9a26788","Type":"ContainerStarted","Data":"47f13e02af6cae4b3052d12938dc83407aa2e55d8a46fbe84971728a429f6203"} Oct 02 14:24:39 crc kubenswrapper[4710]: I1002 14:24:39.137460 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.026549695 podStartE2EDuration="3.137428919s" podCreationTimestamp="2025-10-02 14:24:36 +0000 UTC" firstStartedPulling="2025-10-02 14:24:37.782859144 +0000 UTC m=+5721.889270027" lastFinishedPulling="2025-10-02 14:24:38.893738368 +0000 UTC m=+5723.000149251" observedRunningTime="2025-10-02 14:24:39.134235829 +0000 UTC m=+5723.240646832" watchObservedRunningTime="2025-10-02 14:24:39.137428919 +0000 UTC m=+5723.243839842" Oct 02 14:24:41 crc kubenswrapper[4710]: I1002 14:24:41.685826 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:41 crc kubenswrapper[4710]: I1002 14:24:41.685868 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:41 crc kubenswrapper[4710]: I1002 14:24:41.756503 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:42 crc kubenswrapper[4710]: I1002 14:24:42.211787 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:42 crc kubenswrapper[4710]: I1002 14:24:42.273215 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25bk6"] Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.174125 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-25bk6" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="registry-server" containerID="cri-o://bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe" gracePeriod=2 Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.682353 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.723328 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-utilities\") pod \"08be0527-86e2-47bb-ad8c-189e51c3e49a\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.723380 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs295\" (UniqueName: \"kubernetes.io/projected/08be0527-86e2-47bb-ad8c-189e51c3e49a-kube-api-access-gs295\") pod \"08be0527-86e2-47bb-ad8c-189e51c3e49a\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.723604 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-catalog-content\") pod \"08be0527-86e2-47bb-ad8c-189e51c3e49a\" (UID: \"08be0527-86e2-47bb-ad8c-189e51c3e49a\") " Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.728599 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-utilities" (OuterVolumeSpecName: "utilities") pod "08be0527-86e2-47bb-ad8c-189e51c3e49a" (UID: "08be0527-86e2-47bb-ad8c-189e51c3e49a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.730386 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08be0527-86e2-47bb-ad8c-189e51c3e49a-kube-api-access-gs295" (OuterVolumeSpecName: "kube-api-access-gs295") pod "08be0527-86e2-47bb-ad8c-189e51c3e49a" (UID: "08be0527-86e2-47bb-ad8c-189e51c3e49a"). InnerVolumeSpecName "kube-api-access-gs295". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.751662 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08be0527-86e2-47bb-ad8c-189e51c3e49a" (UID: "08be0527-86e2-47bb-ad8c-189e51c3e49a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.825859 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.826250 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08be0527-86e2-47bb-ad8c-189e51c3e49a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:44 crc kubenswrapper[4710]: I1002 14:24:44.826263 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs295\" (UniqueName: \"kubernetes.io/projected/08be0527-86e2-47bb-ad8c-189e51c3e49a-kube-api-access-gs295\") on node \"crc\" DevicePath \"\"" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.191203 4710 generic.go:334] "Generic (PLEG): container finished" podID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerID="bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe" exitCode=0 Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.191258 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25bk6" event={"ID":"08be0527-86e2-47bb-ad8c-189e51c3e49a","Type":"ContainerDied","Data":"bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe"} Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.191301 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25bk6" event={"ID":"08be0527-86e2-47bb-ad8c-189e51c3e49a","Type":"ContainerDied","Data":"2d85bc0a50bf66bef0558cac5c174592bfd22e3cafa2b582723ac2d7241f2beb"} Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.191364 4710 scope.go:117] "RemoveContainer" containerID="bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.191998 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25bk6" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.241616 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25bk6"] Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.243695 4710 scope.go:117] "RemoveContainer" containerID="6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.263609 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-25bk6"] Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.274481 4710 scope.go:117] "RemoveContainer" containerID="d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.354338 4710 scope.go:117] "RemoveContainer" containerID="bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe" Oct 02 14:24:45 crc kubenswrapper[4710]: E1002 14:24:45.356974 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe\": container with ID starting with bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe not found: ID does not exist" containerID="bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.357019 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe"} err="failed to get container status \"bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe\": rpc error: code = NotFound desc = could not find container \"bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe\": container with ID starting with bc6dc99f7c03e39a32a8b702cdc597cad3cea69969cc26149cc11181709f5cbe not found: ID does not exist" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.357043 4710 scope.go:117] "RemoveContainer" containerID="6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3" Oct 02 14:24:45 crc kubenswrapper[4710]: E1002 14:24:45.359186 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3\": container with ID starting with 6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3 not found: ID does not exist" containerID="6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.359217 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3"} err="failed to get container status \"6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3\": rpc error: code = NotFound desc = could not find container \"6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3\": container with ID starting with 6e6d94974ca8617ed67499742247da83bd1faa95f43478caff87387edf6586a3 not found: ID does not exist" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.359232 4710 scope.go:117] "RemoveContainer" containerID="d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69" Oct 02 14:24:45 crc kubenswrapper[4710]: E1002 14:24:45.359712 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69\": container with ID starting with d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69 not found: ID does not exist" containerID="d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69" Oct 02 14:24:45 crc kubenswrapper[4710]: I1002 14:24:45.359772 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69"} err="failed to get container status \"d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69\": rpc error: code = NotFound desc = could not find container \"d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69\": container with ID starting with d94240faf29fd040c7186438ddff0c4dfd49634ab60cd18fa638eee2a8676d69 not found: ID does not exist" Oct 02 14:24:46 crc kubenswrapper[4710]: I1002 14:24:46.920240 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" path="/var/lib/kubelet/pods/08be0527-86e2-47bb-ad8c-189e51c3e49a/volumes" Oct 02 14:24:52 crc kubenswrapper[4710]: I1002 14:24:52.530186 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:24:52 crc kubenswrapper[4710]: I1002 14:24:52.530763 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.065679 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgflt/must-gather-k6mcj"] Oct 02 14:25:12 crc kubenswrapper[4710]: E1002 14:25:12.066726 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="extract-utilities" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.066785 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="extract-utilities" Oct 02 14:25:12 crc kubenswrapper[4710]: E1002 14:25:12.066817 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="registry-server" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.066826 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="registry-server" Oct 02 14:25:12 crc kubenswrapper[4710]: E1002 14:25:12.066872 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="extract-content" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.066880 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="extract-content" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.067124 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="08be0527-86e2-47bb-ad8c-189e51c3e49a" containerName="registry-server" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.068249 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.069549 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xgflt"/"kube-root-ca.crt" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.070406 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xgflt"/"openshift-service-ca.crt" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.071442 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xgflt"/"default-dockercfg-xmqsd" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.073785 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xgflt/must-gather-k6mcj"] Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.218942 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-must-gather-output\") pod \"must-gather-k6mcj\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.219061 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbs92\" (UniqueName: \"kubernetes.io/projected/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-kube-api-access-gbs92\") pod \"must-gather-k6mcj\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.320445 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbs92\" (UniqueName: \"kubernetes.io/projected/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-kube-api-access-gbs92\") pod \"must-gather-k6mcj\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.320961 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-must-gather-output\") pod \"must-gather-k6mcj\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.321356 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-must-gather-output\") pod \"must-gather-k6mcj\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.346659 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbs92\" (UniqueName: \"kubernetes.io/projected/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-kube-api-access-gbs92\") pod \"must-gather-k6mcj\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.387574 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.831338 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xgflt/must-gather-k6mcj"] Oct 02 14:25:12 crc kubenswrapper[4710]: W1002 14:25:12.839658 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod956ddf0d_9d06_42ab_b6c5_39915e9e64f6.slice/crio-a10510b10595fe290b17684f073131c594d62d69ed57b254a53634515f7d701d WatchSource:0}: Error finding container a10510b10595fe290b17684f073131c594d62d69ed57b254a53634515f7d701d: Status 404 returned error can't find the container with id a10510b10595fe290b17684f073131c594d62d69ed57b254a53634515f7d701d Oct 02 14:25:12 crc kubenswrapper[4710]: I1002 14:25:12.842832 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 14:25:13 crc kubenswrapper[4710]: I1002 14:25:13.540086 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/must-gather-k6mcj" event={"ID":"956ddf0d-9d06-42ab-b6c5-39915e9e64f6","Type":"ContainerStarted","Data":"a10510b10595fe290b17684f073131c594d62d69ed57b254a53634515f7d701d"} Oct 02 14:25:20 crc kubenswrapper[4710]: I1002 14:25:20.619041 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/must-gather-k6mcj" event={"ID":"956ddf0d-9d06-42ab-b6c5-39915e9e64f6","Type":"ContainerStarted","Data":"c0ac4a71ec383ad6903c3e7dbfd8118adb354d098c092a1020cb98873f8bedde"} Oct 02 14:25:20 crc kubenswrapper[4710]: I1002 14:25:20.619432 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/must-gather-k6mcj" event={"ID":"956ddf0d-9d06-42ab-b6c5-39915e9e64f6","Type":"ContainerStarted","Data":"b782dc2bdfa46bcea520267d134dd9091e3593fb191b71fb5f97fa09642d9cc6"} Oct 02 14:25:20 crc kubenswrapper[4710]: I1002 14:25:20.636395 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xgflt/must-gather-k6mcj" podStartSLOduration=1.888868427 podStartE2EDuration="8.636379597s" podCreationTimestamp="2025-10-02 14:25:12 +0000 UTC" firstStartedPulling="2025-10-02 14:25:12.842587549 +0000 UTC m=+5756.948998432" lastFinishedPulling="2025-10-02 14:25:19.590098679 +0000 UTC m=+5763.696509602" observedRunningTime="2025-10-02 14:25:20.633965896 +0000 UTC m=+5764.740376779" watchObservedRunningTime="2025-10-02 14:25:20.636379597 +0000 UTC m=+5764.742790480" Oct 02 14:25:22 crc kubenswrapper[4710]: I1002 14:25:22.530694 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:25:22 crc kubenswrapper[4710]: I1002 14:25:22.531070 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:25:22 crc kubenswrapper[4710]: I1002 14:25:22.531110 4710 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" Oct 02 14:25:22 crc kubenswrapper[4710]: I1002 14:25:22.531716 4710 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a"} pod="openshift-machine-config-operator/machine-config-daemon-tblbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 14:25:22 crc kubenswrapper[4710]: I1002 14:25:22.531786 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" containerID="cri-o://ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" gracePeriod=600 Oct 02 14:25:22 crc kubenswrapper[4710]: E1002 14:25:22.741760 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:25:23 crc kubenswrapper[4710]: I1002 14:25:23.666237 4710 generic.go:334] "Generic (PLEG): container finished" podID="01947451-6af9-4b65-b74c-87e098437b71" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" exitCode=0 Oct 02 14:25:23 crc kubenswrapper[4710]: I1002 14:25:23.666296 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerDied","Data":"ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a"} Oct 02 14:25:23 crc kubenswrapper[4710]: I1002 14:25:23.667610 4710 scope.go:117] "RemoveContainer" containerID="5ac9ea6b4e21821de52736c4c61c86ba05c2a3f9ecbd59da65453c4a85595777" Oct 02 14:25:23 crc kubenswrapper[4710]: I1002 14:25:23.668338 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:25:23 crc kubenswrapper[4710]: E1002 14:25:23.668641 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:25:23 crc kubenswrapper[4710]: I1002 14:25:23.937296 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgflt/crc-debug-8p8d7"] Oct 02 14:25:23 crc kubenswrapper[4710]: I1002 14:25:23.939416 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.059409 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-host\") pod \"crc-debug-8p8d7\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.059786 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9z9f\" (UniqueName: \"kubernetes.io/projected/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-kube-api-access-q9z9f\") pod \"crc-debug-8p8d7\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.161638 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9z9f\" (UniqueName: \"kubernetes.io/projected/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-kube-api-access-q9z9f\") pod \"crc-debug-8p8d7\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.161830 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-host\") pod \"crc-debug-8p8d7\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.162043 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-host\") pod \"crc-debug-8p8d7\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.180461 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9z9f\" (UniqueName: \"kubernetes.io/projected/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-kube-api-access-q9z9f\") pod \"crc-debug-8p8d7\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.261207 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:25:24 crc kubenswrapper[4710]: I1002 14:25:24.680253 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" event={"ID":"e36c510a-d88a-4c1d-b385-ca34f12d6a7a","Type":"ContainerStarted","Data":"ef0b3f27706a6493dafa8966c3b0f5c43455683d2b6ac9085e6aed3724f2873a"} Oct 02 14:25:35 crc kubenswrapper[4710]: I1002 14:25:35.904419 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:25:35 crc kubenswrapper[4710]: E1002 14:25:35.905243 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:25:37 crc kubenswrapper[4710]: I1002 14:25:37.804455 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" event={"ID":"e36c510a-d88a-4c1d-b385-ca34f12d6a7a","Type":"ContainerStarted","Data":"eead865b06fdeb21f394d8817a02f04e2a6c9058cd51d5d820a6495029738837"} Oct 02 14:25:37 crc kubenswrapper[4710]: I1002 14:25:37.823955 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" podStartSLOduration=1.808426774 podStartE2EDuration="14.823934763s" podCreationTimestamp="2025-10-02 14:25:23 +0000 UTC" firstStartedPulling="2025-10-02 14:25:24.300696829 +0000 UTC m=+5768.407107712" lastFinishedPulling="2025-10-02 14:25:37.316204788 +0000 UTC m=+5781.422615701" observedRunningTime="2025-10-02 14:25:37.815594853 +0000 UTC m=+5781.922005736" watchObservedRunningTime="2025-10-02 14:25:37.823934763 +0000 UTC m=+5781.930345646" Oct 02 14:25:47 crc kubenswrapper[4710]: I1002 14:25:47.904436 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:25:47 crc kubenswrapper[4710]: E1002 14:25:47.905241 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:26:01 crc kubenswrapper[4710]: I1002 14:26:01.904413 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:26:01 crc kubenswrapper[4710]: E1002 14:26:01.906451 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:26:13 crc kubenswrapper[4710]: I1002 14:26:13.904434 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:26:13 crc kubenswrapper[4710]: E1002 14:26:13.905269 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:26:28 crc kubenswrapper[4710]: I1002 14:26:28.904614 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:26:28 crc kubenswrapper[4710]: E1002 14:26:28.905380 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:26:42 crc kubenswrapper[4710]: I1002 14:26:42.786451 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54f666d7d6-zzpk4_ea51b4d4-7910-45a6-8320-a299877df017/barbican-api/0.log" Oct 02 14:26:42 crc kubenswrapper[4710]: I1002 14:26:42.799488 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54f666d7d6-zzpk4_ea51b4d4-7910-45a6-8320-a299877df017/barbican-api-log/0.log" Oct 02 14:26:42 crc kubenswrapper[4710]: I1002 14:26:42.904965 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:26:42 crc kubenswrapper[4710]: E1002 14:26:42.905413 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.054616 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c86c47bf8-5k59q_593ff4da-0e41-460e-9b1c-b8d0d032d492/barbican-keystone-listener/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.115004 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c86c47bf8-5k59q_593ff4da-0e41-460e-9b1c-b8d0d032d492/barbican-keystone-listener-log/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.295986 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f7dcd7d87-c6j6p_ddcb91c7-f839-4756-9714-bf04aaf2382a/barbican-worker/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.323089 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f7dcd7d87-c6j6p_ddcb91c7-f839-4756-9714-bf04aaf2382a/barbican-worker-log/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.532367 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-cfbpn_65b945dd-164b-45ec-9262-f5d7db0d15dc/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.773562 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bd2ecce-524b-4abd-adfb-effdd3cc8204/ceilometer-notification-agent/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.811021 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bd2ecce-524b-4abd-adfb-effdd3cc8204/ceilometer-central-agent/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.850225 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bd2ecce-524b-4abd-adfb-effdd3cc8204/proxy-httpd/0.log" Oct 02 14:26:43 crc kubenswrapper[4710]: I1002 14:26:43.962648 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bd2ecce-524b-4abd-adfb-effdd3cc8204/sg-core/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.171781 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4f909e63-ebbe-4888-ace1-0f9595d8b5cc/cinder-api-log/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.247327 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4f909e63-ebbe-4888-ace1-0f9595d8b5cc/cinder-api/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.417890 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9/cinder-scheduler/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.481405 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2b6aa6e4-4ccc-402a-b568-3d1038f2bdf9/probe/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.643974 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-nf64n_ef9d9d20-85b8-46c0-9696-e05d43c16d39/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.835267 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hcgj4_b1dad802-c615-4903-89dc-3e19785fa2ed/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:44 crc kubenswrapper[4710]: I1002 14:26:44.945179 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mz6v6_34701c08-ad93-44d4-b39c-1d153bb5f4be/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.135954 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-dc799465c-hvxxq_b611b238-2861-4371-bbdc-6cfb79039ca4/init/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.289267 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-dc799465c-hvxxq_b611b238-2861-4371-bbdc-6cfb79039ca4/init/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.447628 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-dc799465c-hvxxq_b611b238-2861-4371-bbdc-6cfb79039ca4/dnsmasq-dns/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.513895 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-wp7b7_74b5366f-d978-41b8-85e6-a9bdb91230ce/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.680560 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_939d6b43-cd24-4648-8134-19f384e171bd/glance-httpd/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.699346 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_939d6b43-cd24-4648-8134-19f384e171bd/glance-log/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.874353 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_73b7cc70-507d-4e65-98c0-820912b73208/glance-httpd/0.log" Oct 02 14:26:45 crc kubenswrapper[4710]: I1002 14:26:45.925335 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_73b7cc70-507d-4e65-98c0-820912b73208/glance-log/0.log" Oct 02 14:26:46 crc kubenswrapper[4710]: I1002 14:26:46.176903 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-58474b8ddd-nzgx9_b8267ea9-4756-45d5-a7bc-8985275cf297/horizon/0.log" Oct 02 14:26:46 crc kubenswrapper[4710]: I1002 14:26:46.338420 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9fcdz_c2e4b935-4f26-4ebb-8c57-2ddcb1ef03ed/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:46 crc kubenswrapper[4710]: I1002 14:26:46.415346 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-t9x9g_3a57743e-4e6c-47f4-8a5f-6a78a4c9b6d9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:46 crc kubenswrapper[4710]: I1002 14:26:46.811545 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-58474b8ddd-nzgx9_b8267ea9-4756-45d5-a7bc-8985275cf297/horizon-log/0.log" Oct 02 14:26:46 crc kubenswrapper[4710]: I1002 14:26:46.831112 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323561-glh7z_4e87cc78-a8e9-4ba8-bfbd-f956aa4d0e56/keystone-cron/0.log" Oct 02 14:26:47 crc kubenswrapper[4710]: I1002 14:26:47.017734 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_40732cb5-9326-4916-91dc-21095ecdbca7/kube-state-metrics/0.log" Oct 02 14:26:47 crc kubenswrapper[4710]: I1002 14:26:47.209382 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7fc95bbdbd-gxqdq_269c359c-bcb5-4a9c-97f2-b66a8ee22449/keystone-api/0.log" Oct 02 14:26:47 crc kubenswrapper[4710]: I1002 14:26:47.314128 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-q7787_113f767b-8234-4bc9-8ca1-03d3dd56d806/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:47 crc kubenswrapper[4710]: I1002 14:26:47.844182 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c89fc49-c8w65_cf6923e6-0cef-430d-bf95-e6d3aa1a30e4/neutron-httpd/0.log" Oct 02 14:26:47 crc kubenswrapper[4710]: I1002 14:26:47.870787 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c89fc49-c8w65_cf6923e6-0cef-430d-bf95-e6d3aa1a30e4/neutron-api/0.log" Oct 02 14:26:47 crc kubenswrapper[4710]: I1002 14:26:47.950994 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wgmqq_cfbcee5a-c80c-4ee3-a832-98c079200bab/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:48 crc kubenswrapper[4710]: I1002 14:26:48.814734 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c3ab8e0c-8d4b-41b2-af0b-24ebf616425e/nova-cell0-conductor-conductor/0.log" Oct 02 14:26:49 crc kubenswrapper[4710]: I1002 14:26:49.451238 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_339f91fa-4bac-4673-a8d3-77135bc34f08/nova-cell1-conductor-conductor/0.log" Oct 02 14:26:49 crc kubenswrapper[4710]: I1002 14:26:49.949383 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8bd1f002-8396-4368-8d3c-432a2590da5c/nova-api-log/0.log" Oct 02 14:26:50 crc kubenswrapper[4710]: I1002 14:26:50.007998 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5fd38591-5a16-4ce9-89a0-8d7b94fd0e22/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 14:26:50 crc kubenswrapper[4710]: I1002 14:26:50.136710 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8bd1f002-8396-4368-8d3c-432a2590da5c/nova-api-api/0.log" Oct 02 14:26:50 crc kubenswrapper[4710]: I1002 14:26:50.345330 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-brx4d_ab353848-8250-4624-adeb-72e97fdfdb30/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:50 crc kubenswrapper[4710]: I1002 14:26:50.458726 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e55e8319-4c74-4b16-b045-a6f37f4a1ab5/nova-metadata-log/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.028494 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_34e05f6e-7f38-4c5f-88fc-31be5ed5a01a/nova-scheduler-scheduler/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.106090 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c5d102b1-889f-487b-be09-725834944ff6/memcached/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.175216 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ada9e73f-75af-4b4a-8e52-af2d5b8c3728/mysql-bootstrap/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.365099 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ada9e73f-75af-4b4a-8e52-af2d5b8c3728/mysql-bootstrap/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.412591 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ada9e73f-75af-4b4a-8e52-af2d5b8c3728/galera/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.647481 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_40dcfd31-21c8-4581-b2f6-889737eea66d/mysql-bootstrap/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.768265 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_40dcfd31-21c8-4581-b2f6-889737eea66d/mysql-bootstrap/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.840030 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_40dcfd31-21c8-4581-b2f6-889737eea66d/galera/0.log" Oct 02 14:26:51 crc kubenswrapper[4710]: I1002 14:26:51.994829 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_71a2a3d5-5a6a-4f31-8df3-f087529d550d/openstackclient/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.152051 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fp7wl_d2d9ea68-a957-4056-867d-18e0535bec60/ovn-controller/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.287552 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j7xs2_7250728d-33a5-4939-8fc3-19fe161d8035/openstack-network-exporter/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.354117 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e55e8319-4c74-4b16-b045-a6f37f4a1ab5/nova-metadata-metadata/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.480879 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jd2lt_a8725f3f-36cb-4074-abc0-44031fb37d60/ovsdb-server-init/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.633425 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jd2lt_a8725f3f-36cb-4074-abc0-44031fb37d60/ovsdb-server-init/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.697058 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jd2lt_a8725f3f-36cb-4074-abc0-44031fb37d60/ovsdb-server/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.879743 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jd2lt_a8725f3f-36cb-4074-abc0-44031fb37d60/ovs-vswitchd/0.log" Oct 02 14:26:52 crc kubenswrapper[4710]: I1002 14:26:52.905696 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hcv8m_6c66d4e3-dc39-4b2d-9adf-c3c69b149982/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:53 crc kubenswrapper[4710]: I1002 14:26:53.145205 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c29623cf-e442-438d-8f92-d16927045438/openstack-network-exporter/0.log" Oct 02 14:26:53 crc kubenswrapper[4710]: I1002 14:26:53.350999 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c29623cf-e442-438d-8f92-d16927045438/ovn-northd/0.log" Oct 02 14:26:53 crc kubenswrapper[4710]: I1002 14:26:53.521517 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cbac60db-a726-44d9-998b-cfca036d65d9/openstack-network-exporter/0.log" Oct 02 14:26:53 crc kubenswrapper[4710]: I1002 14:26:53.616203 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cbac60db-a726-44d9-998b-cfca036d65d9/ovsdbserver-nb/0.log" Oct 02 14:26:53 crc kubenswrapper[4710]: I1002 14:26:53.744898 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7e26518b-c86d-4760-a600-b7e8a6594fab/ovsdbserver-sb/0.log" Oct 02 14:26:53 crc kubenswrapper[4710]: I1002 14:26:53.750810 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7e26518b-c86d-4760-a600-b7e8a6594fab/openstack-network-exporter/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.043552 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8cd8bc89b-fpjvs_ed865d67-a052-4171-8d50-7cd44ab89f7d/placement-api/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.190931 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c72cadc-76e1-4e5e-aff8-dec25f7c3283/init-config-reloader/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.324690 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8cd8bc89b-fpjvs_ed865d67-a052-4171-8d50-7cd44ab89f7d/placement-log/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.493912 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c72cadc-76e1-4e5e-aff8-dec25f7c3283/prometheus/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.516377 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c72cadc-76e1-4e5e-aff8-dec25f7c3283/config-reloader/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.557698 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c72cadc-76e1-4e5e-aff8-dec25f7c3283/thanos-sidecar/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.570423 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_4c72cadc-76e1-4e5e-aff8-dec25f7c3283/init-config-reloader/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.721651 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f/setup-container/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.921236 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_98c2f825-d5c6-47e7-a96d-8cf7227b0a24/setup-container/0.log" Oct 02 14:26:54 crc kubenswrapper[4710]: I1002 14:26:54.995109 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f/rabbitmq/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.001948 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2b79e1d-1ec7-4a4c-8cb7-e38d6d392a5f/setup-container/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.151643 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_98c2f825-d5c6-47e7-a96d-8cf7227b0a24/setup-container/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.188013 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_98c2f825-d5c6-47e7-a96d-8cf7227b0a24/rabbitmq/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.326321 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a5eb635a-718a-4361-a3b6-33d483c1e1b2/setup-container/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.416769 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a5eb635a-718a-4361-a3b6-33d483c1e1b2/setup-container/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.460902 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a5eb635a-718a-4361-a3b6-33d483c1e1b2/rabbitmq/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.531634 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8fwsl_7db5b7fb-18a0-493e-91ba-f0bfbd73f3e5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.628476 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rs846_c8c51310-5db6-40e1-b97d-b0460629caed/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.719578 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-mxg9l_3086d3aa-5da8-4765-87dc-f7a122984390/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.881273 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-w925t_95a95ed4-304b-47b9-9468-819007b0690c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.905116 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:26:55 crc kubenswrapper[4710]: E1002 14:26:55.905414 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:26:55 crc kubenswrapper[4710]: I1002 14:26:55.986681 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mnx5v_40227663-ba0e-4abe-877f-1783781a44a1/ssh-known-hosts-edpm-deployment/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.176935 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-76f964d7cf-dhr2b_407d1688-f160-4ee8-934d-9c76f353e4b9/proxy-server/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.273996 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-76f964d7cf-dhr2b_407d1688-f160-4ee8-934d-9c76f353e4b9/proxy-httpd/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.329543 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-xvqwj_989f6404-4404-4e77-b3bc-f3f27a7e7520/swift-ring-rebalance/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.437403 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/account-auditor/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.485970 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/account-reaper/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.555435 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/account-replicator/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.618091 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/container-auditor/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.645983 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/account-server/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.759617 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/container-replicator/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.844979 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/container-updater/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.849882 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/object-auditor/0.log" Oct 02 14:26:56 crc kubenswrapper[4710]: I1002 14:26:56.856958 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/container-server/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.211759 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/object-expirer/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.221002 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/object-server/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.258726 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/object-updater/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.262630 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/object-replicator/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.380125 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/rsync/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.446663 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_67031d3a-15f2-4b86-92d7-b525a3a8420e/swift-recon-cron/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.508891 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2vb9h_76f7bc0e-1c4c-47a8-b2be-7a8d5f2783e8/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.583445 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bql4d"] Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.587734 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.598599 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bql4d"] Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.708555 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-catalog-content\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.708632 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkt6f\" (UniqueName: \"kubernetes.io/projected/83541afc-786a-4c3c-9a0b-78cbecc52826-kube-api-access-rkt6f\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.708738 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-utilities\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.747889 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_22910690-8789-41f2-bd0a-b56b2c6683a6/tempest-tests-tempest-tests-runner/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.763257 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_49a43af4-9bec-4b33-a75a-77f0c9a26788/test-operator-logs-container/0.log" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.811162 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-utilities\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.811268 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-catalog-content\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.811318 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkt6f\" (UniqueName: \"kubernetes.io/projected/83541afc-786a-4c3c-9a0b-78cbecc52826-kube-api-access-rkt6f\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.812036 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-utilities\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.812249 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-catalog-content\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.845570 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkt6f\" (UniqueName: \"kubernetes.io/projected/83541afc-786a-4c3c-9a0b-78cbecc52826-kube-api-access-rkt6f\") pod \"redhat-operators-bql4d\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:57 crc kubenswrapper[4710]: I1002 14:26:57.909382 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:26:58 crc kubenswrapper[4710]: I1002 14:26:58.141982 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-zrgql_d234b8ec-5b0f-4032-9798-517c5d275c1c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 14:26:58 crc kubenswrapper[4710]: I1002 14:26:58.398456 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bql4d"] Oct 02 14:26:58 crc kubenswrapper[4710]: I1002 14:26:58.585480 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerStarted","Data":"7a33e6ad5d5ab0ea7147f2725d02d37d75fdf063eb1fcd653d2d3dd331718c94"} Oct 02 14:26:59 crc kubenswrapper[4710]: I1002 14:26:59.314404 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_3a69c931-f59d-4751-b560-a4b9f92cd3a2/watcher-applier/0.log" Oct 02 14:26:59 crc kubenswrapper[4710]: I1002 14:26:59.373503 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_f8748218-24af-4a2b-a6de-c7ea409ae6f8/watcher-api-log/0.log" Oct 02 14:26:59 crc kubenswrapper[4710]: I1002 14:26:59.593776 4710 generic.go:334] "Generic (PLEG): container finished" podID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerID="6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce" exitCode=0 Oct 02 14:26:59 crc kubenswrapper[4710]: I1002 14:26:59.593865 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerDied","Data":"6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce"} Oct 02 14:27:01 crc kubenswrapper[4710]: I1002 14:27:01.615267 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerStarted","Data":"031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f"} Oct 02 14:27:03 crc kubenswrapper[4710]: I1002 14:27:03.278951 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_549c5440-af2a-4aa2-8d33-a28dde44a69b/watcher-decision-engine/0.log" Oct 02 14:27:03 crc kubenswrapper[4710]: I1002 14:27:03.563913 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_f8748218-24af-4a2b-a6de-c7ea409ae6f8/watcher-api/0.log" Oct 02 14:27:05 crc kubenswrapper[4710]: I1002 14:27:05.653585 4710 generic.go:334] "Generic (PLEG): container finished" podID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerID="031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f" exitCode=0 Oct 02 14:27:05 crc kubenswrapper[4710]: I1002 14:27:05.653658 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerDied","Data":"031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f"} Oct 02 14:27:06 crc kubenswrapper[4710]: I1002 14:27:06.688431 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerStarted","Data":"0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce"} Oct 02 14:27:06 crc kubenswrapper[4710]: I1002 14:27:06.709257 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bql4d" podStartSLOduration=3.092048662 podStartE2EDuration="9.709235428s" podCreationTimestamp="2025-10-02 14:26:57 +0000 UTC" firstStartedPulling="2025-10-02 14:26:59.595549929 +0000 UTC m=+5863.701960812" lastFinishedPulling="2025-10-02 14:27:06.212736695 +0000 UTC m=+5870.319147578" observedRunningTime="2025-10-02 14:27:06.705740969 +0000 UTC m=+5870.812151852" watchObservedRunningTime="2025-10-02 14:27:06.709235428 +0000 UTC m=+5870.815646311" Oct 02 14:27:07 crc kubenswrapper[4710]: I1002 14:27:07.909527 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:27:07 crc kubenswrapper[4710]: I1002 14:27:07.909790 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:27:08 crc kubenswrapper[4710]: I1002 14:27:08.960081 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bql4d" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="registry-server" probeResult="failure" output=< Oct 02 14:27:08 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 14:27:08 crc kubenswrapper[4710]: > Oct 02 14:27:10 crc kubenswrapper[4710]: I1002 14:27:10.905088 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:27:10 crc kubenswrapper[4710]: E1002 14:27:10.905894 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:27:18 crc kubenswrapper[4710]: I1002 14:27:18.999074 4710 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bql4d" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="registry-server" probeResult="failure" output=< Oct 02 14:27:18 crc kubenswrapper[4710]: timeout: failed to connect service ":50051" within 1s Oct 02 14:27:18 crc kubenswrapper[4710]: > Oct 02 14:27:25 crc kubenswrapper[4710]: I1002 14:27:25.905093 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:27:25 crc kubenswrapper[4710]: E1002 14:27:25.905810 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:27:27 crc kubenswrapper[4710]: I1002 14:27:27.969104 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:27:28 crc kubenswrapper[4710]: I1002 14:27:28.034023 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:27:28 crc kubenswrapper[4710]: I1002 14:27:28.811080 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bql4d"] Oct 02 14:27:29 crc kubenswrapper[4710]: I1002 14:27:29.894859 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bql4d" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="registry-server" containerID="cri-o://0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce" gracePeriod=2 Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.461098 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.606729 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-catalog-content\") pod \"83541afc-786a-4c3c-9a0b-78cbecc52826\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.606965 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkt6f\" (UniqueName: \"kubernetes.io/projected/83541afc-786a-4c3c-9a0b-78cbecc52826-kube-api-access-rkt6f\") pod \"83541afc-786a-4c3c-9a0b-78cbecc52826\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.607115 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-utilities\") pod \"83541afc-786a-4c3c-9a0b-78cbecc52826\" (UID: \"83541afc-786a-4c3c-9a0b-78cbecc52826\") " Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.608880 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-utilities" (OuterVolumeSpecName: "utilities") pod "83541afc-786a-4c3c-9a0b-78cbecc52826" (UID: "83541afc-786a-4c3c-9a0b-78cbecc52826"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.626628 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83541afc-786a-4c3c-9a0b-78cbecc52826-kube-api-access-rkt6f" (OuterVolumeSpecName: "kube-api-access-rkt6f") pod "83541afc-786a-4c3c-9a0b-78cbecc52826" (UID: "83541afc-786a-4c3c-9a0b-78cbecc52826"). InnerVolumeSpecName "kube-api-access-rkt6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.700100 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83541afc-786a-4c3c-9a0b-78cbecc52826" (UID: "83541afc-786a-4c3c-9a0b-78cbecc52826"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.710305 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.710356 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkt6f\" (UniqueName: \"kubernetes.io/projected/83541afc-786a-4c3c-9a0b-78cbecc52826-kube-api-access-rkt6f\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.710369 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83541afc-786a-4c3c-9a0b-78cbecc52826-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.911588 4710 generic.go:334] "Generic (PLEG): container finished" podID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerID="0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce" exitCode=0 Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.911675 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bql4d" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.933847 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerDied","Data":"0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce"} Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.933986 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bql4d" event={"ID":"83541afc-786a-4c3c-9a0b-78cbecc52826","Type":"ContainerDied","Data":"7a33e6ad5d5ab0ea7147f2725d02d37d75fdf063eb1fcd653d2d3dd331718c94"} Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.934020 4710 scope.go:117] "RemoveContainer" containerID="0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce" Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.958619 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bql4d"] Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.969964 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bql4d"] Oct 02 14:27:30 crc kubenswrapper[4710]: I1002 14:27:30.991413 4710 scope.go:117] "RemoveContainer" containerID="031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.021506 4710 scope.go:117] "RemoveContainer" containerID="6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce" Oct 02 14:27:31 crc kubenswrapper[4710]: E1002 14:27:31.023536 4710 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83541afc_786a_4c3c_9a0b_78cbecc52826.slice/crio-7a33e6ad5d5ab0ea7147f2725d02d37d75fdf063eb1fcd653d2d3dd331718c94\": RecentStats: unable to find data in memory cache]" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.104885 4710 scope.go:117] "RemoveContainer" containerID="0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce" Oct 02 14:27:31 crc kubenswrapper[4710]: E1002 14:27:31.105609 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce\": container with ID starting with 0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce not found: ID does not exist" containerID="0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.105642 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce"} err="failed to get container status \"0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce\": rpc error: code = NotFound desc = could not find container \"0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce\": container with ID starting with 0779b6a00e6b791fb9285e65795e2dc455b0e8c5217fabc33843ea8916e369ce not found: ID does not exist" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.105666 4710 scope.go:117] "RemoveContainer" containerID="031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f" Oct 02 14:27:31 crc kubenswrapper[4710]: E1002 14:27:31.106302 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f\": container with ID starting with 031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f not found: ID does not exist" containerID="031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.106353 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f"} err="failed to get container status \"031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f\": rpc error: code = NotFound desc = could not find container \"031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f\": container with ID starting with 031de676c5194d63cecbfb266a21f25f9d3b6f4fe389dea08fd1264b4115bf8f not found: ID does not exist" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.106385 4710 scope.go:117] "RemoveContainer" containerID="6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce" Oct 02 14:27:31 crc kubenswrapper[4710]: E1002 14:27:31.106810 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce\": container with ID starting with 6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce not found: ID does not exist" containerID="6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce" Oct 02 14:27:31 crc kubenswrapper[4710]: I1002 14:27:31.106835 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce"} err="failed to get container status \"6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce\": rpc error: code = NotFound desc = could not find container \"6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce\": container with ID starting with 6bfae5713dcecf01fd8ad886dfe3f72991b83b9a2318c8199117a483a01a10ce not found: ID does not exist" Oct 02 14:27:32 crc kubenswrapper[4710]: I1002 14:27:32.923056 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" path="/var/lib/kubelet/pods/83541afc-786a-4c3c-9a0b-78cbecc52826/volumes" Oct 02 14:27:39 crc kubenswrapper[4710]: I1002 14:27:39.905219 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:27:39 crc kubenswrapper[4710]: E1002 14:27:39.906494 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:27:51 crc kubenswrapper[4710]: I1002 14:27:51.188664 4710 generic.go:334] "Generic (PLEG): container finished" podID="e36c510a-d88a-4c1d-b385-ca34f12d6a7a" containerID="eead865b06fdeb21f394d8817a02f04e2a6c9058cd51d5d820a6495029738837" exitCode=0 Oct 02 14:27:51 crc kubenswrapper[4710]: I1002 14:27:51.188830 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" event={"ID":"e36c510a-d88a-4c1d-b385-ca34f12d6a7a","Type":"ContainerDied","Data":"eead865b06fdeb21f394d8817a02f04e2a6c9058cd51d5d820a6495029738837"} Oct 02 14:27:51 crc kubenswrapper[4710]: I1002 14:27:51.905683 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:27:51 crc kubenswrapper[4710]: E1002 14:27:51.905992 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.344919 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.387527 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgflt/crc-debug-8p8d7"] Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.398742 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgflt/crc-debug-8p8d7"] Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.404613 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-host\") pod \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.404741 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-host" (OuterVolumeSpecName: "host") pod "e36c510a-d88a-4c1d-b385-ca34f12d6a7a" (UID: "e36c510a-d88a-4c1d-b385-ca34f12d6a7a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.404776 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9z9f\" (UniqueName: \"kubernetes.io/projected/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-kube-api-access-q9z9f\") pod \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\" (UID: \"e36c510a-d88a-4c1d-b385-ca34f12d6a7a\") " Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.405257 4710 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-host\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.409958 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-kube-api-access-q9z9f" (OuterVolumeSpecName: "kube-api-access-q9z9f") pod "e36c510a-d88a-4c1d-b385-ca34f12d6a7a" (UID: "e36c510a-d88a-4c1d-b385-ca34f12d6a7a"). InnerVolumeSpecName "kube-api-access-q9z9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.507067 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9z9f\" (UniqueName: \"kubernetes.io/projected/e36c510a-d88a-4c1d-b385-ca34f12d6a7a-kube-api-access-q9z9f\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:52 crc kubenswrapper[4710]: I1002 14:27:52.927894 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e36c510a-d88a-4c1d-b385-ca34f12d6a7a" path="/var/lib/kubelet/pods/e36c510a-d88a-4c1d-b385-ca34f12d6a7a/volumes" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.215068 4710 scope.go:117] "RemoveContainer" containerID="eead865b06fdeb21f394d8817a02f04e2a6c9058cd51d5d820a6495029738837" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.215140 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-8p8d7" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.552189 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgflt/crc-debug-7mkqj"] Oct 02 14:27:53 crc kubenswrapper[4710]: E1002 14:27:53.552772 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="extract-content" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.552799 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="extract-content" Oct 02 14:27:53 crc kubenswrapper[4710]: E1002 14:27:53.552837 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36c510a-d88a-4c1d-b385-ca34f12d6a7a" containerName="container-00" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.552847 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36c510a-d88a-4c1d-b385-ca34f12d6a7a" containerName="container-00" Oct 02 14:27:53 crc kubenswrapper[4710]: E1002 14:27:53.552872 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="extract-utilities" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.552886 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="extract-utilities" Oct 02 14:27:53 crc kubenswrapper[4710]: E1002 14:27:53.552918 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="registry-server" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.552929 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="registry-server" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.553247 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36c510a-d88a-4c1d-b385-ca34f12d6a7a" containerName="container-00" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.553278 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="83541afc-786a-4c3c-9a0b-78cbecc52826" containerName="registry-server" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.554307 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.634630 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0542b044-3c12-4024-899b-2e7fff523a3c-host\") pod \"crc-debug-7mkqj\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.634698 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwxbw\" (UniqueName: \"kubernetes.io/projected/0542b044-3c12-4024-899b-2e7fff523a3c-kube-api-access-fwxbw\") pod \"crc-debug-7mkqj\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.736700 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwxbw\" (UniqueName: \"kubernetes.io/projected/0542b044-3c12-4024-899b-2e7fff523a3c-kube-api-access-fwxbw\") pod \"crc-debug-7mkqj\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.736978 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0542b044-3c12-4024-899b-2e7fff523a3c-host\") pod \"crc-debug-7mkqj\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.737126 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0542b044-3c12-4024-899b-2e7fff523a3c-host\") pod \"crc-debug-7mkqj\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.767101 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwxbw\" (UniqueName: \"kubernetes.io/projected/0542b044-3c12-4024-899b-2e7fff523a3c-kube-api-access-fwxbw\") pod \"crc-debug-7mkqj\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:53 crc kubenswrapper[4710]: I1002 14:27:53.876714 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:54 crc kubenswrapper[4710]: I1002 14:27:54.226251 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" event={"ID":"0542b044-3c12-4024-899b-2e7fff523a3c","Type":"ContainerStarted","Data":"22708e93caa2cf4ccd119c8a2cf4660a35f7409c584124b32bf59490d7f8246c"} Oct 02 14:27:54 crc kubenswrapper[4710]: I1002 14:27:54.226803 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" event={"ID":"0542b044-3c12-4024-899b-2e7fff523a3c","Type":"ContainerStarted","Data":"bdc2172f4911f1517a6b0eaf465619fb7a35d4c9c3dc4533ed233bf581e80c09"} Oct 02 14:27:54 crc kubenswrapper[4710]: I1002 14:27:54.254260 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" podStartSLOduration=1.254237475 podStartE2EDuration="1.254237475s" podCreationTimestamp="2025-10-02 14:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 14:27:54.242162951 +0000 UTC m=+5918.348573924" watchObservedRunningTime="2025-10-02 14:27:54.254237475 +0000 UTC m=+5918.360648368" Oct 02 14:27:55 crc kubenswrapper[4710]: I1002 14:27:55.254866 4710 generic.go:334] "Generic (PLEG): container finished" podID="0542b044-3c12-4024-899b-2e7fff523a3c" containerID="22708e93caa2cf4ccd119c8a2cf4660a35f7409c584124b32bf59490d7f8246c" exitCode=0 Oct 02 14:27:55 crc kubenswrapper[4710]: I1002 14:27:55.254916 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" event={"ID":"0542b044-3c12-4024-899b-2e7fff523a3c","Type":"ContainerDied","Data":"22708e93caa2cf4ccd119c8a2cf4660a35f7409c584124b32bf59490d7f8246c"} Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.361587 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.385105 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0542b044-3c12-4024-899b-2e7fff523a3c-host\") pod \"0542b044-3c12-4024-899b-2e7fff523a3c\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.385152 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwxbw\" (UniqueName: \"kubernetes.io/projected/0542b044-3c12-4024-899b-2e7fff523a3c-kube-api-access-fwxbw\") pod \"0542b044-3c12-4024-899b-2e7fff523a3c\" (UID: \"0542b044-3c12-4024-899b-2e7fff523a3c\") " Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.385462 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0542b044-3c12-4024-899b-2e7fff523a3c-host" (OuterVolumeSpecName: "host") pod "0542b044-3c12-4024-899b-2e7fff523a3c" (UID: "0542b044-3c12-4024-899b-2e7fff523a3c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.390981 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0542b044-3c12-4024-899b-2e7fff523a3c-kube-api-access-fwxbw" (OuterVolumeSpecName: "kube-api-access-fwxbw") pod "0542b044-3c12-4024-899b-2e7fff523a3c" (UID: "0542b044-3c12-4024-899b-2e7fff523a3c"). InnerVolumeSpecName "kube-api-access-fwxbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.486334 4710 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0542b044-3c12-4024-899b-2e7fff523a3c-host\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:56 crc kubenswrapper[4710]: I1002 14:27:56.486374 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwxbw\" (UniqueName: \"kubernetes.io/projected/0542b044-3c12-4024-899b-2e7fff523a3c-kube-api-access-fwxbw\") on node \"crc\" DevicePath \"\"" Oct 02 14:27:57 crc kubenswrapper[4710]: I1002 14:27:57.272486 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" event={"ID":"0542b044-3c12-4024-899b-2e7fff523a3c","Type":"ContainerDied","Data":"bdc2172f4911f1517a6b0eaf465619fb7a35d4c9c3dc4533ed233bf581e80c09"} Oct 02 14:27:57 crc kubenswrapper[4710]: I1002 14:27:57.272815 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdc2172f4911f1517a6b0eaf465619fb7a35d4c9c3dc4533ed233bf581e80c09" Oct 02 14:27:57 crc kubenswrapper[4710]: I1002 14:27:57.272880 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-7mkqj" Oct 02 14:28:03 crc kubenswrapper[4710]: I1002 14:28:03.945060 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgflt/crc-debug-7mkqj"] Oct 02 14:28:03 crc kubenswrapper[4710]: I1002 14:28:03.958341 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgflt/crc-debug-7mkqj"] Oct 02 14:28:04 crc kubenswrapper[4710]: I1002 14:28:04.919681 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0542b044-3c12-4024-899b-2e7fff523a3c" path="/var/lib/kubelet/pods/0542b044-3c12-4024-899b-2e7fff523a3c/volumes" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.159336 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgflt/crc-debug-cwvgw"] Oct 02 14:28:05 crc kubenswrapper[4710]: E1002 14:28:05.160169 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0542b044-3c12-4024-899b-2e7fff523a3c" containerName="container-00" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.160187 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0542b044-3c12-4024-899b-2e7fff523a3c" containerName="container-00" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.160644 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0542b044-3c12-4024-899b-2e7fff523a3c" containerName="container-00" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.162256 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.230553 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/502db718-90f5-4def-b470-e93dced9b332-host\") pod \"crc-debug-cwvgw\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.231107 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlnnm\" (UniqueName: \"kubernetes.io/projected/502db718-90f5-4def-b470-e93dced9b332-kube-api-access-nlnnm\") pod \"crc-debug-cwvgw\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.333302 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/502db718-90f5-4def-b470-e93dced9b332-host\") pod \"crc-debug-cwvgw\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.333430 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlnnm\" (UniqueName: \"kubernetes.io/projected/502db718-90f5-4def-b470-e93dced9b332-kube-api-access-nlnnm\") pod \"crc-debug-cwvgw\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.333492 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/502db718-90f5-4def-b470-e93dced9b332-host\") pod \"crc-debug-cwvgw\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.361048 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlnnm\" (UniqueName: \"kubernetes.io/projected/502db718-90f5-4def-b470-e93dced9b332-kube-api-access-nlnnm\") pod \"crc-debug-cwvgw\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.486403 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:05 crc kubenswrapper[4710]: I1002 14:28:05.904253 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:28:05 crc kubenswrapper[4710]: E1002 14:28:05.904833 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:28:06 crc kubenswrapper[4710]: I1002 14:28:06.372788 4710 generic.go:334] "Generic (PLEG): container finished" podID="502db718-90f5-4def-b470-e93dced9b332" containerID="152de4b01693ce4625ae14c3773de5826235d75c520844642b527979b0658161" exitCode=0 Oct 02 14:28:06 crc kubenswrapper[4710]: I1002 14:28:06.372824 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-cwvgw" event={"ID":"502db718-90f5-4def-b470-e93dced9b332","Type":"ContainerDied","Data":"152de4b01693ce4625ae14c3773de5826235d75c520844642b527979b0658161"} Oct 02 14:28:06 crc kubenswrapper[4710]: I1002 14:28:06.372848 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/crc-debug-cwvgw" event={"ID":"502db718-90f5-4def-b470-e93dced9b332","Type":"ContainerStarted","Data":"8064542e744d19acf6a16fa7af69da6403877f05cca25a116603cac82558aa27"} Oct 02 14:28:06 crc kubenswrapper[4710]: I1002 14:28:06.418676 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgflt/crc-debug-cwvgw"] Oct 02 14:28:06 crc kubenswrapper[4710]: I1002 14:28:06.427024 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgflt/crc-debug-cwvgw"] Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.498677 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.582958 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/502db718-90f5-4def-b470-e93dced9b332-host\") pod \"502db718-90f5-4def-b470-e93dced9b332\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.583038 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/502db718-90f5-4def-b470-e93dced9b332-host" (OuterVolumeSpecName: "host") pod "502db718-90f5-4def-b470-e93dced9b332" (UID: "502db718-90f5-4def-b470-e93dced9b332"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.583160 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlnnm\" (UniqueName: \"kubernetes.io/projected/502db718-90f5-4def-b470-e93dced9b332-kube-api-access-nlnnm\") pod \"502db718-90f5-4def-b470-e93dced9b332\" (UID: \"502db718-90f5-4def-b470-e93dced9b332\") " Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.583824 4710 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/502db718-90f5-4def-b470-e93dced9b332-host\") on node \"crc\" DevicePath \"\"" Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.588848 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502db718-90f5-4def-b470-e93dced9b332-kube-api-access-nlnnm" (OuterVolumeSpecName: "kube-api-access-nlnnm") pod "502db718-90f5-4def-b470-e93dced9b332" (UID: "502db718-90f5-4def-b470-e93dced9b332"). InnerVolumeSpecName "kube-api-access-nlnnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:28:07 crc kubenswrapper[4710]: I1002 14:28:07.685037 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlnnm\" (UniqueName: \"kubernetes.io/projected/502db718-90f5-4def-b470-e93dced9b332-kube-api-access-nlnnm\") on node \"crc\" DevicePath \"\"" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.139995 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/util/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.331412 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/pull/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.332023 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/util/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.340642 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/pull/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.389715 4710 scope.go:117] "RemoveContainer" containerID="152de4b01693ce4625ae14c3773de5826235d75c520844642b527979b0658161" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.389792 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/crc-debug-cwvgw" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.502647 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/util/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.523413 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/extract/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.524181 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9888f54e67ad02f331ea10db80d8000c1af2cd8c12eb2ebb2127d3dcbcl5n49_7986fe9f-1265-4339-ac4c-24fba0067f48/pull/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.674852 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-jwd7s_d1945e74-9d29-4e91-aab0-29ed814373bb/kube-rbac-proxy/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.747623 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-jwd7s_d1945e74-9d29-4e91-aab0-29ed814373bb/manager/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.790477 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-4k5fd_0cceb14e-b353-4c50-9790-46206743cadb/kube-rbac-proxy/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.886893 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-4k5fd_0cceb14e-b353-4c50-9790-46206743cadb/manager/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.916920 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502db718-90f5-4def-b470-e93dced9b332" path="/var/lib/kubelet/pods/502db718-90f5-4def-b470-e93dced9b332/volumes" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.957033 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-nsccl_7c8ef37b-d83a-455a-855d-64f2532c2a82/kube-rbac-proxy/0.log" Oct 02 14:28:08 crc kubenswrapper[4710]: I1002 14:28:08.982635 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-nsccl_7c8ef37b-d83a-455a-855d-64f2532c2a82/manager/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.127256 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-759cb_c7eae467-8d9a-41f5-9d91-133fffed772f/kube-rbac-proxy/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.192929 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-759cb_c7eae467-8d9a-41f5-9d91-133fffed772f/manager/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.294672 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-fzzx7_648199c6-9952-4da8-96c7-ea049a9a1d8b/kube-rbac-proxy/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.298024 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-fzzx7_648199c6-9952-4da8-96c7-ea049a9a1d8b/manager/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.378428 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-b96hl_44e7e977-a2d9-4cd8-8b2c-b28c62b95991/kube-rbac-proxy/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.523529 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-b96hl_44e7e977-a2d9-4cd8-8b2c-b28c62b95991/manager/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.548159 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-glgh4_8c6c2206-348e-4fc0-8b76-4ab14e6bccb4/kube-rbac-proxy/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.681045 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-czjvk_60ef26f7-b128-4e27-a8dc-46bf65589a12/kube-rbac-proxy/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.696861 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-glgh4_8c6c2206-348e-4fc0-8b76-4ab14e6bccb4/manager/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.756454 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-czjvk_60ef26f7-b128-4e27-a8dc-46bf65589a12/manager/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.856671 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-fmkmc_e0b5b89e-972b-4961-982d-0d7b04b3e509/kube-rbac-proxy/0.log" Oct 02 14:28:09 crc kubenswrapper[4710]: I1002 14:28:09.939884 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-fmkmc_e0b5b89e-972b-4961-982d-0d7b04b3e509/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.033437 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-2gszd_37052f75-ada0-4aa2-ba11-295678f5a627/kube-rbac-proxy/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.076334 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-2gszd_37052f75-ada0-4aa2-ba11-295678f5a627/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.146538 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-zwvmz_fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc/kube-rbac-proxy/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.243186 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-zwvmz_fc1a1e1c-0a49-4b8d-96c1-68f0211c43bc/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.367726 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-9ktdn_f5b18430-0eb0-46a6-95cb-79aac8b9f170/kube-rbac-proxy/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.445642 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-9ktdn_f5b18430-0eb0-46a6-95cb-79aac8b9f170/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.575781 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-mbcgh_0d644973-169e-4c16-85a1-62680612b413/kube-rbac-proxy/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.629514 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-mbcgh_0d644973-169e-4c16-85a1-62680612b413/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.699211 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-s2pcr_753c96fc-cfc1-46ff-bd18-f7a41a9b8974/kube-rbac-proxy/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.748248 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-s2pcr_753c96fc-cfc1-46ff-bd18-f7a41a9b8974/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.803283 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-x4scm_ccccaf1d-ffb3-471e-bbc9-2f052164b7be/kube-rbac-proxy/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.873677 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-x4scm_ccccaf1d-ffb3-471e-bbc9-2f052164b7be/manager/0.log" Oct 02 14:28:10 crc kubenswrapper[4710]: I1002 14:28:10.972033 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-95f458bf5-dqjqf_42570ebd-0936-455a-aad9-e8e37ebbe05c/kube-rbac-proxy/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.172756 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56f696895d-565tx_e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c/kube-rbac-proxy/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.340312 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56f696895d-565tx_e8d8adaa-4e4b-4684-ac12-c1ebeb51c26c/operator/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.364555 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f6g5f_925e710b-d243-4055-8133-ad8bb3cd81bd/registry-server/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.633450 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-rh965_8a6436d8-ce16-4549-a0b5-811f677dfd0f/kube-rbac-proxy/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.659944 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-rh965_8a6436d8-ce16-4549-a0b5-811f677dfd0f/manager/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.674183 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-6bwxt_d02bda21-fd1d-4058-9e76-7c8df8a864d3/kube-rbac-proxy/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.881033 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-6bwxt_d02bda21-fd1d-4058-9e76-7c8df8a864d3/manager/0.log" Oct 02 14:28:11 crc kubenswrapper[4710]: I1002 14:28:11.941471 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-p2ksk_90363657-f227-44f6-bc20-213a4b7ab60d/operator/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.087016 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-q8lfk_b4183159-214e-472b-a527-e35c091cd5bf/manager/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.145271 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-q8lfk_b4183159-214e-472b-a527-e35c091cd5bf/kube-rbac-proxy/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.186896 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-z7mcb_624fe0ee-ec13-452d-94bd-b883419e021e/kube-rbac-proxy/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.386873 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-nw8rd_b360de58-b6b9-4e33-810d-574a5c0d1b70/manager/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.402464 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-nw8rd_b360de58-b6b9-4e33-810d-574a5c0d1b70/kube-rbac-proxy/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.408168 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-95f458bf5-dqjqf_42570ebd-0936-455a-aad9-e8e37ebbe05c/manager/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.593892 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d597476bd-88c2w_7fb73401-46c9-4b5b-b6f9-c7359e67c668/kube-rbac-proxy/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.608604 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-z7mcb_624fe0ee-ec13-452d-94bd-b883419e021e/manager/0.log" Oct 02 14:28:12 crc kubenswrapper[4710]: I1002 14:28:12.695120 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d597476bd-88c2w_7fb73401-46c9-4b5b-b6f9-c7359e67c668/manager/0.log" Oct 02 14:28:20 crc kubenswrapper[4710]: I1002 14:28:20.904581 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:28:20 crc kubenswrapper[4710]: E1002 14:28:20.906110 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:28:28 crc kubenswrapper[4710]: I1002 14:28:28.306259 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ff49s_69a7d580-fc27-4989-a013-5a30f603fd82/control-plane-machine-set-operator/0.log" Oct 02 14:28:28 crc kubenswrapper[4710]: I1002 14:28:28.453496 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hq8zf_148b577d-5144-4ff2-bacf-a8a0fad52c61/kube-rbac-proxy/0.log" Oct 02 14:28:28 crc kubenswrapper[4710]: I1002 14:28:28.485842 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hq8zf_148b577d-5144-4ff2-bacf-a8a0fad52c61/machine-api-operator/0.log" Oct 02 14:28:34 crc kubenswrapper[4710]: I1002 14:28:34.904841 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:28:34 crc kubenswrapper[4710]: E1002 14:28:34.905888 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:28:41 crc kubenswrapper[4710]: I1002 14:28:41.013536 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9p576_4f8b7f42-0f95-448d-963c-9cfce5078d2b/cert-manager-controller/0.log" Oct 02 14:28:41 crc kubenswrapper[4710]: I1002 14:28:41.119428 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-wbjnf_4d218b44-c736-49fa-b12a-0c21a5ca361b/cert-manager-cainjector/0.log" Oct 02 14:28:41 crc kubenswrapper[4710]: I1002 14:28:41.199026 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rsnsq_7181be4b-6c46-46e5-8d15-453f1b81ed21/cert-manager-webhook/0.log" Oct 02 14:28:48 crc kubenswrapper[4710]: I1002 14:28:48.905962 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:28:48 crc kubenswrapper[4710]: E1002 14:28:48.906961 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:28:53 crc kubenswrapper[4710]: I1002 14:28:53.573662 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-2m97m_fc6708be-e01d-45d3-84b9-34f09476898f/nmstate-console-plugin/0.log" Oct 02 14:28:53 crc kubenswrapper[4710]: I1002 14:28:53.757257 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-xzv7r_fdd7df24-9aea-4c00-9b0a-424cb370d89f/nmstate-handler/0.log" Oct 02 14:28:53 crc kubenswrapper[4710]: I1002 14:28:53.804894 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-f5lpb_34ee392d-77a8-44d3-9bde-ccbb875628fc/kube-rbac-proxy/0.log" Oct 02 14:28:53 crc kubenswrapper[4710]: I1002 14:28:53.805951 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-f5lpb_34ee392d-77a8-44d3-9bde-ccbb875628fc/nmstate-metrics/0.log" Oct 02 14:28:53 crc kubenswrapper[4710]: I1002 14:28:53.942620 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-9lmgf_95074a87-510d-4d71-947e-b8cc858a29b2/nmstate-operator/0.log" Oct 02 14:28:53 crc kubenswrapper[4710]: I1002 14:28:53.991936 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-dtspq_f680c7e5-dff2-426c-9edd-387bfc613153/nmstate-webhook/0.log" Oct 02 14:28:59 crc kubenswrapper[4710]: I1002 14:28:59.905323 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:28:59 crc kubenswrapper[4710]: E1002 14:28:59.906135 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:29:07 crc kubenswrapper[4710]: I1002 14:29:07.630727 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-4x985_3b1c2aba-9eb9-4851-9bbe-75c4ee93f325/kube-rbac-proxy/0.log" Oct 02 14:29:07 crc kubenswrapper[4710]: I1002 14:29:07.833725 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-4x985_3b1c2aba-9eb9-4851-9bbe-75c4ee93f325/controller/0.log" Oct 02 14:29:07 crc kubenswrapper[4710]: I1002 14:29:07.913663 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-frr-files/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.030384 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-reloader/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.049111 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-frr-files/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.085063 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-metrics/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.107609 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-reloader/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.264421 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-frr-files/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.304427 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-metrics/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.304435 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-metrics/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.313719 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-reloader/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.476650 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-metrics/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.482734 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/controller/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.484954 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-frr-files/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.485029 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/cp-reloader/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.666988 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/kube-rbac-proxy-frr/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.678193 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/kube-rbac-proxy/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.700181 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/frr-metrics/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.904623 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/reloader/0.log" Oct 02 14:29:08 crc kubenswrapper[4710]: I1002 14:29:08.930022 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-wv5f2_6ac0bab7-e1c0-4411-8a35-18fa210385a7/frr-k8s-webhook-server/0.log" Oct 02 14:29:09 crc kubenswrapper[4710]: I1002 14:29:09.152314 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7667858fb4-55crv_eaac425c-61dd-47c7-a02a-56deb08ca106/manager/0.log" Oct 02 14:29:09 crc kubenswrapper[4710]: I1002 14:29:09.352489 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68d8677cb8-7smlw_f0d936b3-7f41-4d19-8af8-18b4aa2f8567/webhook-server/0.log" Oct 02 14:29:09 crc kubenswrapper[4710]: I1002 14:29:09.400280 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wh9nb_73bea163-fba2-4a76-8e88-6a31ed86f8f3/kube-rbac-proxy/0.log" Oct 02 14:29:10 crc kubenswrapper[4710]: I1002 14:29:10.124709 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wh9nb_73bea163-fba2-4a76-8e88-6a31ed86f8f3/speaker/0.log" Oct 02 14:29:10 crc kubenswrapper[4710]: I1002 14:29:10.376043 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-swzvg_bfc4a046-7c49-44e9-8324-3d5b54c57442/frr/0.log" Oct 02 14:29:11 crc kubenswrapper[4710]: I1002 14:29:11.905834 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:29:11 crc kubenswrapper[4710]: E1002 14:29:11.906387 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:29:21 crc kubenswrapper[4710]: I1002 14:29:21.656208 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/util/0.log" Oct 02 14:29:21 crc kubenswrapper[4710]: I1002 14:29:21.855609 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/util/0.log" Oct 02 14:29:21 crc kubenswrapper[4710]: I1002 14:29:21.863151 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/pull/0.log" Oct 02 14:29:21 crc kubenswrapper[4710]: I1002 14:29:21.890549 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/pull/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.022783 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/pull/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.050446 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/extract/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.052580 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2nq4r8_1675361e-20c2-46ab-8d3a-0b28c519fc3d/util/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.182023 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/util/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.334045 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/util/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.343630 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/pull/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.389662 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/pull/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.564005 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/pull/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.600825 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/extract/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.622355 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnwh7k_6cf57a3f-610d-4ee5-b2d9-ed967b7bfbde/util/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.745583 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/extract-utilities/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.885093 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/extract-utilities/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.904702 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:29:22 crc kubenswrapper[4710]: E1002 14:29:22.905036 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.935076 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/extract-content/0.log" Oct 02 14:29:22 crc kubenswrapper[4710]: I1002 14:29:22.962367 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/extract-content/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.108028 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/extract-utilities/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.136399 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/extract-content/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.350476 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/extract-utilities/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.516081 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/extract-content/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.565587 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/extract-content/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.647388 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/extract-utilities/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.797348 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8scqr_cc1a0832-a94c-47c9-853b-beee2115ce82/registry-server/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.816500 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/extract-utilities/0.log" Oct 02 14:29:23 crc kubenswrapper[4710]: I1002 14:29:23.826014 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/extract-content/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.014594 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/util/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.256473 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/pull/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.296942 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/util/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.299408 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/pull/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.505714 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/util/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.530673 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/pull/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.575221 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvx2wh_4505d296-ddac-4be5-8ced-3dfc7e8eae59/extract/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.649358 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hqkkb_87ad1040-7c7d-4265-b428-d95694c6c346/registry-server/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.715549 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z2qt5_a54e807e-bc57-41a3-8298-da6597e3e669/marketplace-operator/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.855585 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/extract-utilities/0.log" Oct 02 14:29:24 crc kubenswrapper[4710]: I1002 14:29:24.990666 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/extract-utilities/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.013299 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/extract-content/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.027066 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/extract-content/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.177392 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/extract-utilities/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.195409 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/extract-content/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.270509 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/extract-utilities/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.404581 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lb6tp_2e0e94c3-5165-4c97-a959-b5a26070fa99/registry-server/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.468919 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/extract-utilities/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.484386 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/extract-content/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.505516 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/extract-content/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.717816 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/extract-content/0.log" Oct 02 14:29:25 crc kubenswrapper[4710]: I1002 14:29:25.738294 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/extract-utilities/0.log" Oct 02 14:29:26 crc kubenswrapper[4710]: I1002 14:29:26.152572 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kxkkw_4416a645-9297-4285-a468-e8741689087b/registry-server/0.log" Oct 02 14:29:35 crc kubenswrapper[4710]: I1002 14:29:35.905270 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:29:35 crc kubenswrapper[4710]: E1002 14:29:35.906254 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:29:38 crc kubenswrapper[4710]: I1002 14:29:38.115763 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-f8rs5_4645af58-f9ab-4b88-a1fc-5a1ef95e330d/prometheus-operator/0.log" Oct 02 14:29:38 crc kubenswrapper[4710]: I1002 14:29:38.257058 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7bc78586dc-mnkr2_d224b622-e6dd-47ce-868e-dfd2fe59d7c8/prometheus-operator-admission-webhook/0.log" Oct 02 14:29:38 crc kubenswrapper[4710]: I1002 14:29:38.299893 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7bc78586dc-z9scn_3df7c825-06b5-480f-920e-47216a812bfd/prometheus-operator-admission-webhook/0.log" Oct 02 14:29:38 crc kubenswrapper[4710]: I1002 14:29:38.447517 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-5jcdd_20de3f43-1fd9-47e0-9fae-d84370233104/operator/0.log" Oct 02 14:29:38 crc kubenswrapper[4710]: I1002 14:29:38.474666 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-q42t7_7d54e231-d091-4532-b5e9-5c2f27045e7e/perses-operator/0.log" Oct 02 14:29:48 crc kubenswrapper[4710]: I1002 14:29:48.905199 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:29:48 crc kubenswrapper[4710]: E1002 14:29:48.906684 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:29:59 crc kubenswrapper[4710]: I1002 14:29:59.904273 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:29:59 crc kubenswrapper[4710]: E1002 14:29:59.905020 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.189359 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s"] Oct 02 14:30:00 crc kubenswrapper[4710]: E1002 14:30:00.190045 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502db718-90f5-4def-b470-e93dced9b332" containerName="container-00" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.190077 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="502db718-90f5-4def-b470-e93dced9b332" containerName="container-00" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.192451 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="502db718-90f5-4def-b470-e93dced9b332" containerName="container-00" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.193525 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.196227 4710 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.196535 4710 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.207827 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s"] Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.302128 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp7p9\" (UniqueName: \"kubernetes.io/projected/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-kube-api-access-fp7p9\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.302199 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-config-volume\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.302289 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-secret-volume\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.404101 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-secret-volume\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.404298 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp7p9\" (UniqueName: \"kubernetes.io/projected/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-kube-api-access-fp7p9\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.404342 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-config-volume\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.405417 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-config-volume\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.409992 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-secret-volume\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.422768 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp7p9\" (UniqueName: \"kubernetes.io/projected/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-kube-api-access-fp7p9\") pod \"collect-profiles-29323590-7894s\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:00 crc kubenswrapper[4710]: I1002 14:30:00.531289 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:01 crc kubenswrapper[4710]: I1002 14:30:00.996510 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s"] Oct 02 14:30:01 crc kubenswrapper[4710]: W1002 14:30:01.003432 4710 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dc47a26_c084_4898_a5a3_a1ddf2d22de1.slice/crio-df317bbfec1f648b9a6d619020216424cea53703cad4dad7c790910e0bd13bad WatchSource:0}: Error finding container df317bbfec1f648b9a6d619020216424cea53703cad4dad7c790910e0bd13bad: Status 404 returned error can't find the container with id df317bbfec1f648b9a6d619020216424cea53703cad4dad7c790910e0bd13bad Oct 02 14:30:01 crc kubenswrapper[4710]: I1002 14:30:01.527321 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" event={"ID":"0dc47a26-c084-4898-a5a3-a1ddf2d22de1","Type":"ContainerStarted","Data":"4a54daf7e708687947752f4183970de82165623c85d2f76f505064dccecaaab1"} Oct 02 14:30:01 crc kubenswrapper[4710]: I1002 14:30:01.527614 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" event={"ID":"0dc47a26-c084-4898-a5a3-a1ddf2d22de1","Type":"ContainerStarted","Data":"df317bbfec1f648b9a6d619020216424cea53703cad4dad7c790910e0bd13bad"} Oct 02 14:30:01 crc kubenswrapper[4710]: I1002 14:30:01.551509 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" podStartSLOduration=1.5514592889999999 podStartE2EDuration="1.551459289s" podCreationTimestamp="2025-10-02 14:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 14:30:01.54275959 +0000 UTC m=+6045.649170473" watchObservedRunningTime="2025-10-02 14:30:01.551459289 +0000 UTC m=+6045.657870172" Oct 02 14:30:02 crc kubenswrapper[4710]: I1002 14:30:02.539418 4710 generic.go:334] "Generic (PLEG): container finished" podID="0dc47a26-c084-4898-a5a3-a1ddf2d22de1" containerID="4a54daf7e708687947752f4183970de82165623c85d2f76f505064dccecaaab1" exitCode=0 Oct 02 14:30:02 crc kubenswrapper[4710]: I1002 14:30:02.539466 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" event={"ID":"0dc47a26-c084-4898-a5a3-a1ddf2d22de1","Type":"ContainerDied","Data":"4a54daf7e708687947752f4183970de82165623c85d2f76f505064dccecaaab1"} Oct 02 14:30:03 crc kubenswrapper[4710]: I1002 14:30:03.909418 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.079608 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-secret-volume\") pod \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.080028 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp7p9\" (UniqueName: \"kubernetes.io/projected/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-kube-api-access-fp7p9\") pod \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.080198 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-config-volume\") pod \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\" (UID: \"0dc47a26-c084-4898-a5a3-a1ddf2d22de1\") " Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.080790 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-config-volume" (OuterVolumeSpecName: "config-volume") pod "0dc47a26-c084-4898-a5a3-a1ddf2d22de1" (UID: "0dc47a26-c084-4898-a5a3-a1ddf2d22de1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.080911 4710 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.086085 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-kube-api-access-fp7p9" (OuterVolumeSpecName: "kube-api-access-fp7p9") pod "0dc47a26-c084-4898-a5a3-a1ddf2d22de1" (UID: "0dc47a26-c084-4898-a5a3-a1ddf2d22de1"). InnerVolumeSpecName "kube-api-access-fp7p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.086128 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0dc47a26-c084-4898-a5a3-a1ddf2d22de1" (UID: "0dc47a26-c084-4898-a5a3-a1ddf2d22de1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.183033 4710 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.183550 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp7p9\" (UniqueName: \"kubernetes.io/projected/0dc47a26-c084-4898-a5a3-a1ddf2d22de1-kube-api-access-fp7p9\") on node \"crc\" DevicePath \"\"" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.562340 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" event={"ID":"0dc47a26-c084-4898-a5a3-a1ddf2d22de1","Type":"ContainerDied","Data":"df317bbfec1f648b9a6d619020216424cea53703cad4dad7c790910e0bd13bad"} Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.562376 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323590-7894s" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.562733 4710 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df317bbfec1f648b9a6d619020216424cea53703cad4dad7c790910e0bd13bad" Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.652690 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77"] Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.661900 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323545-b8k77"] Oct 02 14:30:04 crc kubenswrapper[4710]: I1002 14:30:04.922777 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ad83ee3-894e-4059-89c7-b1605e4a4b65" path="/var/lib/kubelet/pods/3ad83ee3-894e-4059-89c7-b1605e4a4b65/volumes" Oct 02 14:30:10 crc kubenswrapper[4710]: I1002 14:30:10.905544 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:30:10 crc kubenswrapper[4710]: E1002 14:30:10.906132 4710 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tblbf_openshift-machine-config-operator(01947451-6af9-4b65-b74c-87e098437b71)\"" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" Oct 02 14:30:24 crc kubenswrapper[4710]: I1002 14:30:24.904004 4710 scope.go:117] "RemoveContainer" containerID="ecf22d6a151e6a34514738327fb48a1502f2c18621d0ea7461df70873f2e111a" Oct 02 14:30:25 crc kubenswrapper[4710]: I1002 14:30:25.853607 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" event={"ID":"01947451-6af9-4b65-b74c-87e098437b71","Type":"ContainerStarted","Data":"47be19a0b31f3fadee39c4d1e3c80a505bdd98cb265d23d2eb041b3d5eccf3c8"} Oct 02 14:30:49 crc kubenswrapper[4710]: I1002 14:30:49.406420 4710 scope.go:117] "RemoveContainer" containerID="7baca12dbfbf17eab225dab9032783139c35b155a94dff89ab7757f3840fabf2" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.894063 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r5skn"] Oct 02 14:31:17 crc kubenswrapper[4710]: E1002 14:31:17.895163 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc47a26-c084-4898-a5a3-a1ddf2d22de1" containerName="collect-profiles" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.895176 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc47a26-c084-4898-a5a3-a1ddf2d22de1" containerName="collect-profiles" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.895422 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dc47a26-c084-4898-a5a3-a1ddf2d22de1" containerName="collect-profiles" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.897090 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.929839 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5skn"] Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.939052 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8rd8\" (UniqueName: \"kubernetes.io/projected/a3782cad-0299-4355-ba66-48d698902e31-kube-api-access-w8rd8\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.939120 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-catalog-content\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:17 crc kubenswrapper[4710]: I1002 14:31:17.939353 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-utilities\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.041038 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8rd8\" (UniqueName: \"kubernetes.io/projected/a3782cad-0299-4355-ba66-48d698902e31-kube-api-access-w8rd8\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.041089 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-catalog-content\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.041217 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-utilities\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.041588 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-catalog-content\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.041612 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-utilities\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.062530 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8rd8\" (UniqueName: \"kubernetes.io/projected/a3782cad-0299-4355-ba66-48d698902e31-kube-api-access-w8rd8\") pod \"certified-operators-r5skn\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.223804 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:18 crc kubenswrapper[4710]: I1002 14:31:18.557290 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5skn"] Oct 02 14:31:19 crc kubenswrapper[4710]: I1002 14:31:19.491822 4710 generic.go:334] "Generic (PLEG): container finished" podID="a3782cad-0299-4355-ba66-48d698902e31" containerID="37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839" exitCode=0 Oct 02 14:31:19 crc kubenswrapper[4710]: I1002 14:31:19.491902 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerDied","Data":"37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839"} Oct 02 14:31:19 crc kubenswrapper[4710]: I1002 14:31:19.492182 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerStarted","Data":"a0ee01fae5446c9f64389546183969cc8a4de1d852a538cee91557c086d37c0b"} Oct 02 14:31:19 crc kubenswrapper[4710]: I1002 14:31:19.494734 4710 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 14:31:21 crc kubenswrapper[4710]: I1002 14:31:21.517636 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerStarted","Data":"6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798"} Oct 02 14:31:22 crc kubenswrapper[4710]: I1002 14:31:22.533849 4710 generic.go:334] "Generic (PLEG): container finished" podID="a3782cad-0299-4355-ba66-48d698902e31" containerID="6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798" exitCode=0 Oct 02 14:31:22 crc kubenswrapper[4710]: I1002 14:31:22.533900 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerDied","Data":"6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798"} Oct 02 14:31:23 crc kubenswrapper[4710]: I1002 14:31:23.545642 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerStarted","Data":"51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47"} Oct 02 14:31:23 crc kubenswrapper[4710]: I1002 14:31:23.565272 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r5skn" podStartSLOduration=3.030413534 podStartE2EDuration="6.565248968s" podCreationTimestamp="2025-10-02 14:31:17 +0000 UTC" firstStartedPulling="2025-10-02 14:31:19.494318637 +0000 UTC m=+6123.600729530" lastFinishedPulling="2025-10-02 14:31:23.029154061 +0000 UTC m=+6127.135564964" observedRunningTime="2025-10-02 14:31:23.559795391 +0000 UTC m=+6127.666206274" watchObservedRunningTime="2025-10-02 14:31:23.565248968 +0000 UTC m=+6127.671659861" Oct 02 14:31:28 crc kubenswrapper[4710]: I1002 14:31:28.224837 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:28 crc kubenswrapper[4710]: I1002 14:31:28.225488 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:28 crc kubenswrapper[4710]: I1002 14:31:28.290469 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:28 crc kubenswrapper[4710]: I1002 14:31:28.673367 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:28 crc kubenswrapper[4710]: I1002 14:31:28.734150 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5skn"] Oct 02 14:31:30 crc kubenswrapper[4710]: I1002 14:31:30.620842 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r5skn" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="registry-server" containerID="cri-o://51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47" gracePeriod=2 Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.162438 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.317864 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-utilities\") pod \"a3782cad-0299-4355-ba66-48d698902e31\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.317987 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8rd8\" (UniqueName: \"kubernetes.io/projected/a3782cad-0299-4355-ba66-48d698902e31-kube-api-access-w8rd8\") pod \"a3782cad-0299-4355-ba66-48d698902e31\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.318180 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-catalog-content\") pod \"a3782cad-0299-4355-ba66-48d698902e31\" (UID: \"a3782cad-0299-4355-ba66-48d698902e31\") " Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.319032 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-utilities" (OuterVolumeSpecName: "utilities") pod "a3782cad-0299-4355-ba66-48d698902e31" (UID: "a3782cad-0299-4355-ba66-48d698902e31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.333004 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3782cad-0299-4355-ba66-48d698902e31-kube-api-access-w8rd8" (OuterVolumeSpecName: "kube-api-access-w8rd8") pod "a3782cad-0299-4355-ba66-48d698902e31" (UID: "a3782cad-0299-4355-ba66-48d698902e31"). InnerVolumeSpecName "kube-api-access-w8rd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.385024 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3782cad-0299-4355-ba66-48d698902e31" (UID: "a3782cad-0299-4355-ba66-48d698902e31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.420747 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.420827 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3782cad-0299-4355-ba66-48d698902e31-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.420848 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8rd8\" (UniqueName: \"kubernetes.io/projected/a3782cad-0299-4355-ba66-48d698902e31-kube-api-access-w8rd8\") on node \"crc\" DevicePath \"\"" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.648616 4710 generic.go:334] "Generic (PLEG): container finished" podID="a3782cad-0299-4355-ba66-48d698902e31" containerID="51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47" exitCode=0 Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.648689 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerDied","Data":"51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47"} Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.648706 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5skn" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.648754 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5skn" event={"ID":"a3782cad-0299-4355-ba66-48d698902e31","Type":"ContainerDied","Data":"a0ee01fae5446c9f64389546183969cc8a4de1d852a538cee91557c086d37c0b"} Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.648797 4710 scope.go:117] "RemoveContainer" containerID="51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.702661 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5skn"] Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.706712 4710 scope.go:117] "RemoveContainer" containerID="6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.713925 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r5skn"] Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.735645 4710 scope.go:117] "RemoveContainer" containerID="37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.790179 4710 scope.go:117] "RemoveContainer" containerID="51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47" Oct 02 14:31:31 crc kubenswrapper[4710]: E1002 14:31:31.790668 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47\": container with ID starting with 51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47 not found: ID does not exist" containerID="51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.790716 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47"} err="failed to get container status \"51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47\": rpc error: code = NotFound desc = could not find container \"51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47\": container with ID starting with 51d97a7a459a62a7b18270ac4906fab0d61b34847537e75fa805669372688a47 not found: ID does not exist" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.790745 4710 scope.go:117] "RemoveContainer" containerID="6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798" Oct 02 14:31:31 crc kubenswrapper[4710]: E1002 14:31:31.791024 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798\": container with ID starting with 6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798 not found: ID does not exist" containerID="6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.791057 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798"} err="failed to get container status \"6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798\": rpc error: code = NotFound desc = could not find container \"6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798\": container with ID starting with 6b45e560171d16a16e7b4752b18b8f9fb89b80007e17fbcd454adcf9c3447798 not found: ID does not exist" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.791077 4710 scope.go:117] "RemoveContainer" containerID="37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839" Oct 02 14:31:31 crc kubenswrapper[4710]: E1002 14:31:31.791333 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839\": container with ID starting with 37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839 not found: ID does not exist" containerID="37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839" Oct 02 14:31:31 crc kubenswrapper[4710]: I1002 14:31:31.791361 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839"} err="failed to get container status \"37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839\": rpc error: code = NotFound desc = could not find container \"37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839\": container with ID starting with 37f9f9ccacdc787f3306cb19586739d6d3872f1558f54a67851c050fe63bd839 not found: ID does not exist" Oct 02 14:31:32 crc kubenswrapper[4710]: I1002 14:31:32.926261 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3782cad-0299-4355-ba66-48d698902e31" path="/var/lib/kubelet/pods/a3782cad-0299-4355-ba66-48d698902e31/volumes" Oct 02 14:31:51 crc kubenswrapper[4710]: I1002 14:31:51.892229 4710 generic.go:334] "Generic (PLEG): container finished" podID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerID="b782dc2bdfa46bcea520267d134dd9091e3593fb191b71fb5f97fa09642d9cc6" exitCode=0 Oct 02 14:31:51 crc kubenswrapper[4710]: I1002 14:31:51.892362 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgflt/must-gather-k6mcj" event={"ID":"956ddf0d-9d06-42ab-b6c5-39915e9e64f6","Type":"ContainerDied","Data":"b782dc2bdfa46bcea520267d134dd9091e3593fb191b71fb5f97fa09642d9cc6"} Oct 02 14:31:51 crc kubenswrapper[4710]: I1002 14:31:51.893303 4710 scope.go:117] "RemoveContainer" containerID="b782dc2bdfa46bcea520267d134dd9091e3593fb191b71fb5f97fa09642d9cc6" Oct 02 14:31:52 crc kubenswrapper[4710]: I1002 14:31:52.761135 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgflt_must-gather-k6mcj_956ddf0d-9d06-42ab-b6c5-39915e9e64f6/gather/0.log" Oct 02 14:32:00 crc kubenswrapper[4710]: I1002 14:32:00.846666 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgflt/must-gather-k6mcj"] Oct 02 14:32:00 crc kubenswrapper[4710]: I1002 14:32:00.847547 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xgflt/must-gather-k6mcj" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="copy" containerID="cri-o://c0ac4a71ec383ad6903c3e7dbfd8118adb354d098c092a1020cb98873f8bedde" gracePeriod=2 Oct 02 14:32:00 crc kubenswrapper[4710]: I1002 14:32:00.854562 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgflt/must-gather-k6mcj"] Oct 02 14:32:00 crc kubenswrapper[4710]: I1002 14:32:00.990445 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgflt_must-gather-k6mcj_956ddf0d-9d06-42ab-b6c5-39915e9e64f6/copy/0.log" Oct 02 14:32:00 crc kubenswrapper[4710]: I1002 14:32:00.991463 4710 generic.go:334] "Generic (PLEG): container finished" podID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerID="c0ac4a71ec383ad6903c3e7dbfd8118adb354d098c092a1020cb98873f8bedde" exitCode=143 Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.321038 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgflt_must-gather-k6mcj_956ddf0d-9d06-42ab-b6c5-39915e9e64f6/copy/0.log" Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.321546 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.378585 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-must-gather-output\") pod \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.378764 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbs92\" (UniqueName: \"kubernetes.io/projected/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-kube-api-access-gbs92\") pod \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\" (UID: \"956ddf0d-9d06-42ab-b6c5-39915e9e64f6\") " Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.389124 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-kube-api-access-gbs92" (OuterVolumeSpecName: "kube-api-access-gbs92") pod "956ddf0d-9d06-42ab-b6c5-39915e9e64f6" (UID: "956ddf0d-9d06-42ab-b6c5-39915e9e64f6"). InnerVolumeSpecName "kube-api-access-gbs92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.481628 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbs92\" (UniqueName: \"kubernetes.io/projected/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-kube-api-access-gbs92\") on node \"crc\" DevicePath \"\"" Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.580421 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "956ddf0d-9d06-42ab-b6c5-39915e9e64f6" (UID: "956ddf0d-9d06-42ab-b6c5-39915e9e64f6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:32:01 crc kubenswrapper[4710]: I1002 14:32:01.582930 4710 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/956ddf0d-9d06-42ab-b6c5-39915e9e64f6-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 14:32:02 crc kubenswrapper[4710]: I1002 14:32:02.002736 4710 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgflt_must-gather-k6mcj_956ddf0d-9d06-42ab-b6c5-39915e9e64f6/copy/0.log" Oct 02 14:32:02 crc kubenswrapper[4710]: I1002 14:32:02.003449 4710 scope.go:117] "RemoveContainer" containerID="c0ac4a71ec383ad6903c3e7dbfd8118adb354d098c092a1020cb98873f8bedde" Oct 02 14:32:02 crc kubenswrapper[4710]: I1002 14:32:02.003472 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgflt/must-gather-k6mcj" Oct 02 14:32:02 crc kubenswrapper[4710]: I1002 14:32:02.026523 4710 scope.go:117] "RemoveContainer" containerID="b782dc2bdfa46bcea520267d134dd9091e3593fb191b71fb5f97fa09642d9cc6" Oct 02 14:32:02 crc kubenswrapper[4710]: I1002 14:32:02.919553 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" path="/var/lib/kubelet/pods/956ddf0d-9d06-42ab-b6c5-39915e9e64f6/volumes" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.692726 4710 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-brvkz"] Oct 02 14:32:38 crc kubenswrapper[4710]: E1002 14:32:38.693831 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="registry-server" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.693849 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="registry-server" Oct 02 14:32:38 crc kubenswrapper[4710]: E1002 14:32:38.693867 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="copy" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.693875 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="copy" Oct 02 14:32:38 crc kubenswrapper[4710]: E1002 14:32:38.693889 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="extract-content" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.693899 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="extract-content" Oct 02 14:32:38 crc kubenswrapper[4710]: E1002 14:32:38.693922 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="extract-utilities" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.693930 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="extract-utilities" Oct 02 14:32:38 crc kubenswrapper[4710]: E1002 14:32:38.693959 4710 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="gather" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.693966 4710 state_mem.go:107] "Deleted CPUSet assignment" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="gather" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.694202 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3782cad-0299-4355-ba66-48d698902e31" containerName="registry-server" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.694229 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="gather" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.694243 4710 memory_manager.go:354] "RemoveStaleState removing state" podUID="956ddf0d-9d06-42ab-b6c5-39915e9e64f6" containerName="copy" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.696246 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.715925 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brvkz"] Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.845231 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-utilities\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.845293 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktx8p\" (UniqueName: \"kubernetes.io/projected/17d322e1-deee-4c9a-8e47-bff64acf8b9d-kube-api-access-ktx8p\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.845588 4710 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-catalog-content\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.947422 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-utilities\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.947480 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktx8p\" (UniqueName: \"kubernetes.io/projected/17d322e1-deee-4c9a-8e47-bff64acf8b9d-kube-api-access-ktx8p\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.947545 4710 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-catalog-content\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.948014 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-catalog-content\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.948033 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-utilities\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:38 crc kubenswrapper[4710]: I1002 14:32:38.966714 4710 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktx8p\" (UniqueName: \"kubernetes.io/projected/17d322e1-deee-4c9a-8e47-bff64acf8b9d-kube-api-access-ktx8p\") pod \"community-operators-brvkz\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:39 crc kubenswrapper[4710]: I1002 14:32:39.036430 4710 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:39 crc kubenswrapper[4710]: I1002 14:32:39.553149 4710 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brvkz"] Oct 02 14:32:40 crc kubenswrapper[4710]: I1002 14:32:40.482984 4710 generic.go:334] "Generic (PLEG): container finished" podID="17d322e1-deee-4c9a-8e47-bff64acf8b9d" containerID="352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d" exitCode=0 Oct 02 14:32:40 crc kubenswrapper[4710]: I1002 14:32:40.483026 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerDied","Data":"352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d"} Oct 02 14:32:40 crc kubenswrapper[4710]: I1002 14:32:40.483246 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerStarted","Data":"145548fe6347664901c77538e5de3f78ee2a5b961f4da10064aadebd82d9fdec"} Oct 02 14:32:42 crc kubenswrapper[4710]: I1002 14:32:42.508808 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerStarted","Data":"424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4"} Oct 02 14:32:43 crc kubenswrapper[4710]: I1002 14:32:43.524563 4710 generic.go:334] "Generic (PLEG): container finished" podID="17d322e1-deee-4c9a-8e47-bff64acf8b9d" containerID="424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4" exitCode=0 Oct 02 14:32:43 crc kubenswrapper[4710]: I1002 14:32:43.524648 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerDied","Data":"424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4"} Oct 02 14:32:44 crc kubenswrapper[4710]: I1002 14:32:44.538617 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerStarted","Data":"8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593"} Oct 02 14:32:44 crc kubenswrapper[4710]: I1002 14:32:44.574511 4710 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-brvkz" podStartSLOduration=3.098432976 podStartE2EDuration="6.574484452s" podCreationTimestamp="2025-10-02 14:32:38 +0000 UTC" firstStartedPulling="2025-10-02 14:32:40.48601223 +0000 UTC m=+6204.592423103" lastFinishedPulling="2025-10-02 14:32:43.962063656 +0000 UTC m=+6208.068474579" observedRunningTime="2025-10-02 14:32:44.569603909 +0000 UTC m=+6208.676014832" watchObservedRunningTime="2025-10-02 14:32:44.574484452 +0000 UTC m=+6208.680895345" Oct 02 14:32:49 crc kubenswrapper[4710]: I1002 14:32:49.036959 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:49 crc kubenswrapper[4710]: I1002 14:32:49.037838 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:49 crc kubenswrapper[4710]: I1002 14:32:49.113520 4710 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:49 crc kubenswrapper[4710]: I1002 14:32:49.658879 4710 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:49 crc kubenswrapper[4710]: I1002 14:32:49.734578 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brvkz"] Oct 02 14:32:51 crc kubenswrapper[4710]: I1002 14:32:51.625037 4710 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-brvkz" podUID="17d322e1-deee-4c9a-8e47-bff64acf8b9d" containerName="registry-server" containerID="cri-o://8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593" gracePeriod=2 Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.193843 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.284157 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-catalog-content\") pod \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.284436 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktx8p\" (UniqueName: \"kubernetes.io/projected/17d322e1-deee-4c9a-8e47-bff64acf8b9d-kube-api-access-ktx8p\") pod \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.284479 4710 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-utilities\") pod \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\" (UID: \"17d322e1-deee-4c9a-8e47-bff64acf8b9d\") " Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.285652 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-utilities" (OuterVolumeSpecName: "utilities") pod "17d322e1-deee-4c9a-8e47-bff64acf8b9d" (UID: "17d322e1-deee-4c9a-8e47-bff64acf8b9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.291322 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d322e1-deee-4c9a-8e47-bff64acf8b9d-kube-api-access-ktx8p" (OuterVolumeSpecName: "kube-api-access-ktx8p") pod "17d322e1-deee-4c9a-8e47-bff64acf8b9d" (UID: "17d322e1-deee-4c9a-8e47-bff64acf8b9d"). InnerVolumeSpecName "kube-api-access-ktx8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.352762 4710 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17d322e1-deee-4c9a-8e47-bff64acf8b9d" (UID: "17d322e1-deee-4c9a-8e47-bff64acf8b9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.386459 4710 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktx8p\" (UniqueName: \"kubernetes.io/projected/17d322e1-deee-4c9a-8e47-bff64acf8b9d-kube-api-access-ktx8p\") on node \"crc\" DevicePath \"\"" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.386522 4710 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.386540 4710 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d322e1-deee-4c9a-8e47-bff64acf8b9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.530102 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.530170 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.647416 4710 generic.go:334] "Generic (PLEG): container finished" podID="17d322e1-deee-4c9a-8e47-bff64acf8b9d" containerID="8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593" exitCode=0 Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.647466 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerDied","Data":"8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593"} Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.647481 4710 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brvkz" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.647502 4710 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvkz" event={"ID":"17d322e1-deee-4c9a-8e47-bff64acf8b9d","Type":"ContainerDied","Data":"145548fe6347664901c77538e5de3f78ee2a5b961f4da10064aadebd82d9fdec"} Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.647526 4710 scope.go:117] "RemoveContainer" containerID="8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.681464 4710 scope.go:117] "RemoveContainer" containerID="424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.694236 4710 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brvkz"] Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.705797 4710 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-brvkz"] Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.714675 4710 scope.go:117] "RemoveContainer" containerID="352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.768147 4710 scope.go:117] "RemoveContainer" containerID="8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593" Oct 02 14:32:52 crc kubenswrapper[4710]: E1002 14:32:52.768616 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593\": container with ID starting with 8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593 not found: ID does not exist" containerID="8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.768668 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593"} err="failed to get container status \"8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593\": rpc error: code = NotFound desc = could not find container \"8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593\": container with ID starting with 8dbbb5da622e194f2a49a1926c4dd3ab4a5a213bc07a5c8c245d27adc436b593 not found: ID does not exist" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.768703 4710 scope.go:117] "RemoveContainer" containerID="424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4" Oct 02 14:32:52 crc kubenswrapper[4710]: E1002 14:32:52.769020 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4\": container with ID starting with 424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4 not found: ID does not exist" containerID="424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.769051 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4"} err="failed to get container status \"424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4\": rpc error: code = NotFound desc = could not find container \"424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4\": container with ID starting with 424d08c9d56239d28d441d5ae0289e5e0897c48bc9175150e0e9e7326dbb08c4 not found: ID does not exist" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.769070 4710 scope.go:117] "RemoveContainer" containerID="352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d" Oct 02 14:32:52 crc kubenswrapper[4710]: E1002 14:32:52.769234 4710 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d\": container with ID starting with 352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d not found: ID does not exist" containerID="352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.769270 4710 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d"} err="failed to get container status \"352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d\": rpc error: code = NotFound desc = could not find container \"352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d\": container with ID starting with 352944ae3b8999c0858f071e93457d8f9c5aa1f1ab5d2f15dede67e06fccfd1d not found: ID does not exist" Oct 02 14:32:52 crc kubenswrapper[4710]: I1002 14:32:52.921127 4710 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d322e1-deee-4c9a-8e47-bff64acf8b9d" path="/var/lib/kubelet/pods/17d322e1-deee-4c9a-8e47-bff64acf8b9d/volumes" Oct 02 14:33:22 crc kubenswrapper[4710]: I1002 14:33:22.530820 4710 patch_prober.go:28] interesting pod/machine-config-daemon-tblbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 14:33:22 crc kubenswrapper[4710]: I1002 14:33:22.531433 4710 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tblbf" podUID="01947451-6af9-4b65-b74c-87e098437b71" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067506524024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067506525017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067471702016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067471703015467 5ustar corecore